diff --git a/.github/workflows/e2e-matrix.yml b/.github/workflows/e2e-matrix.yml new file mode 100644 index 0000000000..cbeb0c3db4 --- /dev/null +++ b/.github/workflows/e2e-matrix.yml @@ -0,0 +1,358 @@ +# Bootstrap no-op workflow to enable PR runs for future e2e matrix + +name: E2E Matrix Tests (bootstrap) + +on: + pull_request: + types: [opened, reopened, synchronize, labeled, unlabeled] + branches: + - main + - feat/ci-e2e-matrix + workflow_dispatch: + +permissions: + contents: read + +jobs: + noop: + name: Bootstrap + runs-on: ubuntu-latest + steps: + - name: Say hello + run: echo "Bootstrap workflow OK" + +# Copyright 2025 Flant JSC +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +name: E2E Matrix Tests (DVP-over-DVP) + +on: + pull_request: + types: [opened, reopened, synchronize, labeled, unlabeled] + branches: + - main + - feat/ci-e2e-matrix + schedule: + - cron: "30 2 * * *" + workflow_dispatch: + inputs: + profiles: + description: "Storage profiles (comma-separated): sds, cephrbd" + required: false + default: "sds,cephrbd" + timeout: + description: "Ginkgo timeout (e.g. 2h, 4h)" + required: false + default: "4h" + +permissions: + contents: read + +env: + E2E_K8S_URL: https://api.e2e.virtlab.flant.com + +jobs: + # ============================================ + # 1. SETUP - Environment preparation + # ============================================ + setup: + name: Setup Environment + runs-on: ubuntu-latest + outputs: + profiles: ${{ steps.load.outputs.profiles }} + steps: + - uses: actions/checkout@v4 + + - name: Load storage profiles + id: load + run: | + # For now, use hardcoded profiles. Later this can be made dynamic + PROFILES='["sds", "cephrbd"]' + echo "profiles=$PROFILES" >> "$GITHUB_OUTPUT" + + - name: Print matrix + run: | + echo "Will test profiles: ${{ steps.load.outputs.profiles }}" + + # ============================================ + # 2. VALIDATE - Configuration validation + # ============================================ + validate: + name: Validate Configuration + needs: setup + runs-on: ubuntu-latest + steps: + - uses: actions/checkout@v4 + + - name: Validate YAML syntax + run: | + echo "✓ Checking YAML syntax..." + python3 -c "import yaml; yaml.safe_load(open('ci/dvp-e2e/values.yaml')); print('✅ values.yaml YAML syntax is valid')" + python3 -c "import yaml; yaml.safe_load(open('.github/workflows/e2e-matrix.yml')); print('✅ e2e-matrix.yml YAML syntax is valid')" + + - name: Check required secrets + run: | + echo "✓ Checking required secrets..." + if [ -z "${{ secrets.E2E_VIRTUALIZATION_SA_SECRET }}" ]; then + echo "❌ E2E_VIRTUALIZATION_SA_SECRET secret is required" + exit 1 + fi + echo "✅ Required secrets are present" + echo "✅ E2E_K8S_URL is set as workflow variable: ${{ vars.E2E_K8S_URL }}" + + # ============================================ + # 3. E2E - Parallel test execution + # ============================================ + e2e: + name: E2E (${{ matrix.profile }}) + needs: [setup, validate] + runs-on: ubuntu-latest + timeout-minutes: 300 + concurrency: + group: e2e-${{ github.ref }}-${{ matrix.profile }} + cancel-in-progress: false + strategy: + fail-fast: false + matrix: + profile: ${{ fromJson(needs.setup.outputs.profiles) }} + + env: + GO_VERSION: "1.24.6" + TMP_ROOT: ${{ github.workspace }}/ci/dvp-e2e/tmp + LOOP_WEBHOOK: ${{ secrets.LOOP_WEBHOOK_URL || secrets.LOOP_WEBHOOK }} + LOOP_CHANNEL: ${{ secrets.LOOP_CHANNEL || 'test-virtualization-loop-alerts' }} # TODO: replace with channel secret after successful run + + steps: + - uses: actions/checkout@v4 + + - name: Set up Go + uses: actions/setup-go@v5 + with: + go-version: ${{ env.GO_VERSION }} + + - uses: ./.github/actions/setup-e2e-tools + + - name: Prepare environment + id: prep + run: | + RUN_ID="nightly-${{ matrix.profile }}-${{ github.run_id }}" + echo "run_id=$RUN_ID" >> "$GITHUB_OUTPUT" + echo "RUN_ID=$RUN_ID" >> "$GITHUB_ENV" + echo "PROFILE=${{ matrix.profile }}" >> "$GITHUB_ENV" + echo "TMP_ROOT=${{ env.TMP_ROOT }}" >> "$GITHUB_ENV" + mkdir -p "${{ env.TMP_ROOT }}/shared" "${{ env.TMP_ROOT }}/matrix-logs" + + - name: Build parent kubeconfig + shell: bash + run: | + if [ -n "${{ vars.E2E_K8S_URL }}" ] && [ -n "${{ secrets.E2E_VIRTUALIZATION_SA_SECRET }}" ]; then + KCFG='${{ env.TMP_ROOT }}/shared/parent-admin.conf' + TOKEN="$(echo '${{ secrets.E2E_VIRTUALIZATION_SA_SECRET }}' | base64 -d 2>/dev/null || echo '${{ secrets.E2E_VIRTUALIZATION_SA_SECRET }}')" + printf '%s\n' \ + 'apiVersion: v1' \ + 'kind: Config' \ + 'clusters:' \ + '- cluster:' \ + " server: ${{ vars.E2E_K8S_URL }}" \ + ' insecure-skip-tls-verify: true' \ + ' name: parent' \ + 'contexts:' \ + '- context:' \ + ' cluster: parent' \ + ' user: sa' \ + ' name: parent' \ + 'current-context: parent' \ + 'users:' \ + '- name: sa' \ + ' user:' \ + " token: ${TOKEN}" > "$KCFG" + chmod 600 "$KCFG" + echo "PARENT_KUBECONFIG_FILE=$KCFG" >> "$GITHUB_ENV" + else + echo "E2E_K8S_URL/E2E_VIRTUALIZATION_SA_SECRET not set" >&2 + exit 1 + fi + + - name: Run E2E tests + working-directory: ci/dvp-e2e + env: + KUBECONFIG: ${{ env.PARENT_KUBECONFIG_FILE }} + REGISTRY_DOCKER_CFG: ${{ secrets.REGISTRY_DOCKER_CFG }} + TMP_ROOT: ${{ env.TMP_ROOT }} + PARENT_KUBECONFIG_FILE: ${{ env.PARENT_KUBECONFIG_FILE }} + E2E_DIR: ${{ github.workspace }}/tests/e2e + run: | + ./bin/e2e-runner \ + --run-id "${RUN_ID}" \ + --timeout "${{ inputs.timeout || '4h' }}" \ + --junit "../../artifacts/${RUN_ID}/junit.xml" \ + ${{ matrix.profile }} + + - name: Upload test logs + if: always() + uses: actions/upload-artifact@v4 + with: + name: logs-${{ matrix.profile }}-${{ steps.prep.outputs.run_id }} + path: ci/dvp-e2e/tmp/matrix-logs/*.log + if-no-files-found: warn + + - name: Upload JUnit report + if: always() + uses: actions/upload-artifact@v4 + with: + name: junit-${{ matrix.profile }}-${{ steps.prep.outputs.run_id }} + path: ci/dvp-e2e/artifacts/**/junit.xml + if-no-files-found: warn + + # ============================================ + # 4. REPORT - Result aggregation + # ============================================ + report: + name: Report Results + needs: e2e + if: always() + runs-on: ubuntu-latest + steps: + - uses: actions/checkout@v4 + + - name: Download all JUnit reports + uses: actions/download-artifact@v4 + with: + pattern: junit-* + path: ./results + + - name: Send individual reports to Loop + working-directory: ci/dvp-e2e + env: + LOOP_WEBHOOK: ${{ secrets.LOOP_WEBHOOK_URL || secrets.LOOP_WEBHOOK }} + run: | + if [ -z "${LOOP_WEBHOOK}" ]; then + echo "No LOOP_WEBHOOK configured; skipping report sending." + exit 0 + fi + + # Send individual reports for each profile + for junit_file in ../../results/junit-*/junit.xml; do + if [ -f "$junit_file" ]; then + profile=$(echo "$junit_file" | sed 's/.*junit-\([^-]*\)-.*/\1/') + run_id=$(echo "$junit_file" | sed 's/.*-\([^-]*\)\/junit.xml/\1/') + task loop:junit:parse \ + JUNIT_FILE="$junit_file" \ + RUN_ID="$run_id" \ + STORAGE_PROFILE="$profile" \ + TEST_TIMEOUT="${{ inputs.timeout || '4h' }}" + fi + done + + - name: Generate matrix summary + if: always() + working-directory: ci/dvp-e2e + env: + KUBECONFIG: ${{ env.PARENT_KUBECONFIG_FILE }} + run: | + python3 scripts/loop_matrix_summary.py \ + --profiles "sds,cephrbd" \ + --run-id-prefix "nightly" \ + --log-dir "tmp/matrix-logs" \ + --webhook-url "${{ secrets.LOOP_WEBHOOK_URL || secrets.LOOP_WEBHOOK }}" \ + --channel "${{ secrets.LOOP_CHANNEL || 'test-virtualization-loop-alerts' }}" > matrix_summary.md || true + DATE=$(date +"%Y-%m-%d") + HASH=$(head -c 16 /dev/urandom | base64 | tr -dc 'a-zA-Z0-9' | head -c 8) + kubectl apply -f - <> $GITHUB_STEP_SUMMARY + echo "" >> $GITHUB_STEP_SUMMARY + echo "| Profile | Status |" >> $GITHUB_STEP_SUMMARY + echo "|---------|--------|" >> $GITHUB_STEP_SUMMARY + + # Check each profile result + for profile in sds cephrbd; do + if [ -f "./results/junit-${profile}-*/junit.xml" ]; then + echo "| $profile | ✅ Completed |" >> $GITHUB_STEP_SUMMARY + else + echo "| $profile | ❌ Failed/Missing |" >> $GITHUB_STEP_SUMMARY + fi + done + + # ============================================ + # 5. CLEANUP - Resource cleanup + # ============================================ + cleanup: + name: Cleanup Resources + needs: e2e + if: always() + runs-on: ubuntu-latest + steps: + - uses: actions/checkout@v4 + + - uses: ./.github/actions/setup-e2e-tools + + - name: Setup kubeconfig for cleanup + run: | + if [ -n "${{ vars.E2E_K8S_URL }}" ] && [ -n "${{ secrets.E2E_VIRTUALIZATION_SA_SECRET }}" ]; then + mkdir -p ~/.kube + TOKEN="$(echo '${{ secrets.E2E_VIRTUALIZATION_SA_SECRET }}' | base64 -d 2>/dev/null || echo '${{ secrets.E2E_VIRTUALIZATION_SA_SECRET }}')" + printf '%s\n' \ + 'apiVersion: v1' \ + 'kind: Config' \ + 'clusters:' \ + '- cluster:' \ + " server: ${{ vars.E2E_K8S_URL }}" \ + ' insecure-skip-tls-verify: true' \ + ' name: parent' \ + 'contexts:' \ + '- context:' \ + ' cluster: parent' \ + ' user: sa' \ + ' name: parent' \ + 'current-context: parent' \ + 'users:' \ + '- name: sa' \ + ' user:' \ + " token: ${TOKEN}" > ~/.kube/config + chmod 600 ~/.kube/config + else + echo "⚠️ Cannot setup kubeconfig for cleanup - secrets not available" + exit 0 + fi + + - name: Cleanup test namespaces + working-directory: ci/dvp-e2e + run: | + echo "🧹 Cleaning up test namespaces..." + task cleanup:namespaces:safe \ + FILTER_PREFIX="nightly-" \ + CONFIRM=true || echo "⚠️ Cleanup completed with warnings" + + - name: Report cleanup results + if: always() + run: | + echo "### Cleanup Results" >> $GITHUB_STEP_SUMMARY + echo "✅ Cleanup job completed" >> $GITHUB_STEP_SUMMARY + echo "🧹 Attempted to clean up namespaces matching 'nightly-*'" >> $GITHUB_STEP_SUMMARY diff --git a/artifacts/auto-clean-sds-20251009-094100-1934/junit.xml b/artifacts/auto-clean-sds-20251009-094100-1934/junit.xml new file mode 100644 index 0000000000..d8a633a96f --- /dev/null +++ b/artifacts/auto-clean-sds-20251009-094100-1934/junit.xml @@ -0,0 +1,2 @@ + + \ No newline at end of file diff --git a/artifacts/clean-cephrbd-20251012-114926-22431/junit.xml b/artifacts/clean-cephrbd-20251012-114926-22431/junit.xml new file mode 100644 index 0000000000..fadffaf299 --- /dev/null +++ b/artifacts/clean-cephrbd-20251012-114926-22431/junit.xml @@ -0,0 +1,830 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 12:14:52.106 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 12:14:56.818 (4.712s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 12:14:56.818 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 12:14:56.818 (0s) + + + > Enter [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/12/25 12:14:56.822 < Exit [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/12/25 12:14:58.855 (2.034s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/12/25 12:14:58.855 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/12/25 12:15:05.203 (6.348s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/12/25 12:15:05.203 < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/12/25 12:15:05.204 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:77 @ 10/12/25 12:31:47.372 + > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/12/25 12:15:05.204 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:76 @ 10/12/25 12:15:05.204 [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:77 @ 10/12/25 12:31:47.372 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/12/25 12:31:47.372 (16m42.151s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/12/25 12:31:47.372 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/12/25 12:31:53.503 (6.13s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:86 @ 10/12/25 12:31:53.503 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:100 @ 10/12/25 12:31:53.503 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:116 @ 10/12/25 12:31:53.503 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:131 @ 10/12/25 12:31:53.503 + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 @ 10/12/25 12:48:42.784 + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/12/25 12:31:53.503 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/12/25 12:31:55.477 (1.973s) > Enter [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/12/25 12:31:55.477 < Exit [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/12/25 12:32:00.642 (5.165s) > Enter [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/12/25 12:32:00.642 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 @ 10/12/25 12:32:00.642 [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 @ 10/12/25 12:48:42.784 < Exit [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/12/25 12:48:42.784 (16m42.127s) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/12/25 12:48:42.784 The list of pods is empty; nothing to dump. STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 12:48:48.776 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 12:48:53.97 (5.193s) < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/12/25 12:48:53.97 (11.185s) + + + + > Enter [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/12/25 12:48:53.97 [SKIPPED] Module SDN is disabled. Skipping all tests for module SDN. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:50 @ 10/12/25 12:48:54.552 < Exit [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/12/25 12:48:54.552 (582ms) > Enter [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/12/25 12:48:54.552 < Exit [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/12/25 12:48:54.552 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/12/25 12:48:54.553 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/12/25 12:48:54.553 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:111 @ 10/12/25 12:48:54.553 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:125 @ 10/12/25 12:48:54.553 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:141 @ 10/12/25 12:48:54.553 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:156 @ 10/12/25 12:48:54.554 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:169 @ 10/12/25 12:48:54.554 + + + + > Enter [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/12/25 12:48:54.554 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:41 @ 10/12/25 12:48:54.554 < Exit [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/12/25 12:48:54.554 (0s) > Enter [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/12/25 12:48:54.554 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 12:48:54.554 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 12:48:59.188 (4.634s) < Exit [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/12/25 12:48:59.188 (4.634s) + + + > Enter [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/12/25 12:48:59.189 < Exit [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/12/25 12:49:01.186 (1.997s) > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 12:49:01.186 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 12:49:01.186 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/12/25 12:49:01.187 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/12/25 12:49:08.429 (7.242s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 12:49:08.429 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 12:49:08.429 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 12:49:08.429 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 12:49:08.429 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/12/25 12:49:08.429 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:91 @ 10/12/25 12:49:08.429 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/12/25 12:49:12.694 (4.265s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 12:49:12.694 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 12:49:12.694 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-not-existing-vmclass-with-changing\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:103 @ 10/12/25 13:05:54.799 + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 12:49:12.695 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 12:49:12.695 (0s) > Enter [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/12/25 12:49:12.695 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:102 @ 10/12/25 12:49:12.695 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-not-existing-vmclass-with-changing\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:103 @ 10/12/25 13:05:54.799 < Exit [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/12/25 13:05:54.8 (16m42.09s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 13:05:54.8 The list of pods is empty; nothing to dump. < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 13:06:01.318 (6.518s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/12/25 13:06:01.318 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:126 @ 10/12/25 13:06:01.318 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:140 @ 10/12/25 13:06:01.318 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:152 @ 10/12/25 13:06:01.319 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:157 @ 10/12/25 13:06:01.319 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:163 @ 10/12/25 13:06:01.319 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:176 @ 10/12/25 13:06:01.319 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:181 @ 10/12/25 13:06:01.319 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:187 @ 10/12/25 13:06:01.319 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:202 @ 10/12/25 13:06:01.319 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:216 @ 10/12/25 13:06:01.319 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:240 @ 10/12/25 13:06:01.319 + + + [FAILED] Warning: resource namespaces/head-5073ae15-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "ceph-pool-r2-csi-rbd" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x1400030a820>: exit status 1 { ProcessState: { pid: 62683, status: 256, rusage: { Utime: { Sec: 0, Usec: 300234, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 57750, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 70303744, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5108, Majflt: 4, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 49, Msgrcv: 56, Nsignals: 364, Nvcsw: 376, Nivcsw: 2645, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/12/25 13:06:08.645 + > Enter [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/12/25 13:06:01.319 < Exit [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/12/25 13:06:04.434 (3.115s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/12/25 13:06:04.434 [FAILED] Warning: resource namespaces/head-5073ae15-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "ceph-pool-r2-csi-rbd" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x1400030a820>: exit status 1 { ProcessState: { pid: 62683, status: 256, rusage: { Utime: { Sec: 0, Usec: 300234, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 57750, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 70303744, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5108, Majflt: 4, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 49, Msgrcv: 56, Nsignals: 364, Nvcsw: 376, Nivcsw: 2645, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/12/25 13:06:08.645 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/12/25 13:06:08.646 (4.211s) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/12/25 13:06:08.646 The list of pods is empty; nothing to dump. < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/12/25 13:06:14.494 (5.849s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:112 @ 10/12/25 13:06:14.495 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:147 @ 10/12/25 13:06:14.495 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:190 @ 10/12/25 13:06:14.495 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:196 @ 10/12/25 13:06:14.495 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:204 @ 10/12/25 13:06:14.495 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:224 @ 10/12/25 13:06:14.495 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:236 @ 10/12/25 13:06:14.496 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:258 @ 10/12/25 13:06:14.496 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:284 @ 10/12/25 13:06:14.496 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:294 @ 10/12/25 13:06:14.496 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:309 @ 10/12/25 13:06:14.496 + + + > Enter [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/12/25 13:06:14.496 < Exit [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/12/25 13:06:17.068 (2.572s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/12/25 13:06:17.069 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/12/25 13:06:22.283 (5.214s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/12/25 13:06:22.283 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/12/25 13:06:22.283 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/12/25 13:06:22.283 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:89 @ 10/12/25 13:06:22.283 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/12/25 13:06:28.151 (5.868s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/12/25 13:06:28.151 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/12/25 13:06:28.151 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:4>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-manual-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-manual-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:100 @ 10/12/25 13:23:10.215 + > Enter [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/12/25 13:06:28.151 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:4>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-manual-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-manual-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:100 @ 10/12/25 13:23:10.215 < Exit [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/12/25 13:23:10.217 (16m42.082s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/12/25 13:23:10.217 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/12/25 13:23:16.401 (6.184s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:109 @ 10/12/25 13:23:16.402 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:123 @ 10/12/25 13:23:16.402 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:147 @ 10/12/25 13:23:16.402 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:161 @ 10/12/25 13:23:16.402 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:183 @ 10/12/25 13:23:16.402 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:202 @ 10/12/25 13:23:16.402 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:226 @ 10/12/25 13:23:16.402 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:240 @ 10/12/25 13:23:16.402 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:250 @ 10/12/25 13:23:16.402 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:265 @ 10/12/25 13:23:16.402 + + + > Enter [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/12/25 13:23:16.402 < Exit [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/12/25 13:23:18.371 (1.969s) > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/12/25 13:23:18.371 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/12/25 13:23:18.371 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/12/25 13:23:18.371 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/12/25 13:23:24.691 (6.32s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/12/25 13:23:24.691 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/12/25 13:23:24.692 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-always-on-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-always-on-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 @ 10/12/25 13:40:06.777 + > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/12/25 13:23:24.692 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/12/25 13:23:24.692 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/12/25 13:23:24.692 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/12/25 13:23:24.692 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/12/25 13:40:06.777 (16m42.098s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-always-on-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-always-on-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 @ 10/12/25 13:40:06.777 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/12/25 13:40:06.778 (16m42.1s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/12/25 13:40:06.778 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/12/25 13:40:13.002 (6.223s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:131 @ 10/12/25 13:40:13.002 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:177 @ 10/12/25 13:40:13.002 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:341 @ 10/12/25 13:40:13.002 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:367 @ 10/12/25 13:40:13.002 + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.002 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 13:40:13.003 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.003 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.003 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 13:40:13.003 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.003 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.003 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 13:40:13.003 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.003 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.003 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 13:40:13.004 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.004 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.004 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 13:40:13.004 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.004 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.004 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 13:40:13.004 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.004 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.004 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 13:40:13.004 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.004 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.005 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 13:40:13.005 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.005 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.005 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 13:40:13.005 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 13:40:13.005 (0s) + + + > Enter [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/12/25 13:40:13.005 < Exit [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/12/25 13:40:15.006 (2.001s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/12/25 13:40:15.006 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/12/25 13:40:23.952 (8.946s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/12/25 13:40:23.952 < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/12/25 13:40:23.952 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:6, cap:6>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-d --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-d\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-b-not-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-b-not-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-c-and-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-c-and-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-affinity --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-affinity\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-selector --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-selector\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-a-not-b --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-a-not-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:103 @ 10/12/25 13:57:15.567 + > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/12/25 13:40:23.952 STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/12/25 13:40:23.952 END STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/12/25 13:40:30.773 (6.82s) STEP: `VirtualMachineClasses` should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:95 @ 10/12/25 13:40:30.773 END STEP: `VirtualMachineClasses` should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:95 @ 10/12/25 13:40:33.504 (2.731s) STEP: `VirtualDisks` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:102 @ 10/12/25 13:40:33.504 END STEP: `VirtualDisks` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:102 @ 10/12/25 13:57:15.567 (16m42.122s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:6, cap:6>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-d --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-d\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-b-not-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-b-not-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-c-and-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-c-and-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-affinity --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-affinity\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-selector --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-selector\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-a-not-b --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-a-not-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:103 @ 10/12/25 13:57:15.567 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/12/25 13:57:15.568 (16m51.674s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/12/25 13:57:15.568 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/12/25 13:57:21.708 (6.141s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:120 @ 10/12/25 13:57:21.709 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:283 @ 10/12/25 13:57:21.709 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:372 @ 10/12/25 13:57:21.709 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:467 @ 10/12/25 13:57:21.709 + + + + > Enter [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/12/25 13:57:21.709 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:73 @ 10/12/25 13:57:21.71 < Exit [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/12/25 13:57:21.71 (0s) > Enter [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/12/25 13:57:21.71 < Exit [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/12/25 13:57:21.71 (0s) > Enter [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/12/25 13:57:21.71 < Exit [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/12/25 13:57:21.71 (0s) + + + > Enter [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/12/25 13:57:21.71 < Exit [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/12/25 13:57:23.687 (1.977s) > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 13:57:23.687 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 13:57:23.687 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/12/25 13:57:23.687 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/12/25 13:57:25.966 (2.279s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 13:57:25.966 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 13:57:25.966 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 13:57:25.966 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 13:57:25.966 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/12/25 13:57:25.966 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:68 @ 10/12/25 13:57:25.967 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/12/25 13:57:44.189 (18.222s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 13:57:44.189 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 13:57:44.189 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 13:57:44.189 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 13:57:44.189 (0s) > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/12/25 13:57:44.189 STEP: VM should be in Running phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:79 @ 10/12/25 13:57:44.189 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/12/25 13:57:47.213 (3.024s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 13:57:47.213 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 13:57:47.213 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 13:57:47.213 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 13:57:47.213 (0s) > Enter [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/12/25 13:57:47.213 < Exit [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/12/25 13:57:47.213 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 13:57:47.213 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 13:57:47.213 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 13:57:47.213 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 13:57:47.213 (0s) > Enter [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/12/25 13:57:47.213 < Exit [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/12/25 13:57:47.213 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 13:57:47.213 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 13:57:47.213 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 13:57:47.213 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 13:57:47.213 (0s) > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/12/25 13:57:47.213 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 13:57:47.213 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 13:58:58.624 (1m11.412s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/12/25 13:58:58.624 (1m11.412s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 13:58:58.624 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 13:58:58.624 (0s) + + + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/12/25 13:58:58.624 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/12/25 13:59:00.626 (2.002s) > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 13:59:00.626 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 13:59:00.626 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/12/25 13:59:00.626 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/12/25 13:59:04.004 (3.378s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 13:59:04.004 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 13:59:04.004 (0s) + + + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 13:59:04.004 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 13:59:04.004 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/12/25 13:59:04.004 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:88 @ 10/12/25 13:59:04.004 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/12/25 13:59:09.362 (5.358s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 13:59:09.362 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 13:59:09.362 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-with-hotplug-standalone --namespace head-5073ae15-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/12/25 14:15:51.446 + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 13:59:09.363 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 13:59:09.363 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/12/25 13:59:09.363 STEP: VDs with consumers should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:99 @ 10/12/25 13:59:09.363 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-with-hotplug-standalone --namespace head-5073ae15-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/12/25 14:15:51.446 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/12/25 14:15:51.448 (16m42.098s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 14:15:51.448 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 14:15:57.552 (6.104s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/12/25 14:15:57.552 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/12/25 14:15:57.552 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/12/25 14:15:57.552 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/12/25 14:15:57.552 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/12/25 14:15:57.552 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/12/25 14:15:57.552 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/12/25 14:15:57.552 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/12/25 14:15:57.552 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/12/25 14:15:57.552 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/12/25 14:15:57.552 + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:51 @ 10/12/25 14:15:59.502 + > Enter [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/12/25 14:15:57.553 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:51 @ 10/12/25 14:15:59.502 < Exit [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/12/25 14:15:59.503 (1.95s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/12/25 14:15:59.503 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/12/25 14:16:05.633 (6.13s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:94 @ 10/12/25 14:16:05.633 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:103 @ 10/12/25 14:16:05.633 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:114 @ 10/12/25 14:16:05.633 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:128 @ 10/12/25 14:16:05.633 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:139 @ 10/12/25 14:16:05.633 + + + + > Enter [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/12/25 14:16:05.633 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:51 @ 10/12/25 14:16:05.633 < Exit [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/12/25 14:16:05.633 (0s) > Enter [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/12/25 14:16:05.634 < Exit [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/12/25 14:16:05.634 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:82 @ 10/12/25 14:16:05.634 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:94 @ 10/12/25 14:16:05.634 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:106 @ 10/12/25 14:16:05.634 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:116 @ 10/12/25 14:16:05.634 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:132 @ 10/12/25 14:16:05.634 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:151 @ 10/12/25 14:16:05.634 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:157 @ 10/12/25 14:16:05.634 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:189 @ 10/12/25 14:16:05.634 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:216 @ 10/12/25 14:16:05.634 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:252 @ 10/12/25 14:16:05.634 + + + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/12/25 14:16:05.634 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/12/25 14:16:07.594 (1.96s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/12/25 14:16:07.594 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/12/25 14:16:11.765 (4.171s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 14:16:11.765 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 14:16:11.765 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/12/25 14:16:11.765 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:97 @ 10/12/25 14:16:11.765 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/12/25 14:16:23.642 (11.877s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 14:16:23.642 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 14:16:23.642 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-b --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-a --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 @ 10/12/25 14:33:05.827 + > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/12/25 14:16:23.643 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 @ 10/12/25 14:16:23.643 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-b --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-a --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 @ 10/12/25 14:33:05.827 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/12/25 14:33:05.829 (16m42.139s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 14:33:05.829 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 14:33:12.009 (6.18s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:118 @ 10/12/25 14:33:12.009 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:129 @ 10/12/25 14:33:12.009 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:142 @ 10/12/25 14:33:12.009 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:166 @ 10/12/25 14:33:12.009 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:174 @ 10/12/25 14:33:12.009 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:179 @ 10/12/25 14:33:12.009 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:187 @ 10/12/25 14:33:12.009 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:198 @ 10/12/25 14:33:12.009 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:209 @ 10/12/25 14:33:12.01 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:222 @ 10/12/25 14:33:12.01 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:228 @ 10/12/25 14:33:12.01 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:240 @ 10/12/25 14:33:12.01 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:250 @ 10/12/25 14:33:12.01 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:258 @ 10/12/25 14:33:12.01 + + + > Enter [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/12/25 14:33:12.01 < Exit [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/12/25 14:33:12.01 (0s) > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:33:12.01 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:33:12.01 (0s) > Enter [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/12/25 14:33:12.01 < Exit [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/12/25 14:33:13.222 (1.212s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:33:13.222 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:33:13.222 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:33:13.222 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:33:13.222 (0s) > Enter [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/12/25 14:33:13.222 STEP: Project should be deployed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:73 @ 10/12/25 14:33:13.222 < Exit [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/12/25 14:33:15.91 (2.687s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:33:15.91 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:33:15.91 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:33:15.91 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:33:15.91 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/12/25 14:33:15.91 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/12/25 14:33:19.088 (3.178s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:33:19.088 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:33:19.088 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:33:19.088 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:33:19.088 (0s) > Enter [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/12/25 14:33:19.088 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/12/25 14:33:19.088 < Exit [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/12/25 14:33:25.265 (6.177s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:33:25.265 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:33:25.266 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-cvi --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-cvi\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/12/25 14:50:07.465 + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:33:25.266 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:33:25.266 (0s) > Enter [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/12/25 14:33:25.266 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/12/25 14:33:25.266 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-cvi --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-cvi\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/12/25 14:50:07.465 < Exit [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/12/25 14:50:07.466 (16m42.231s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:50:07.466 The list of pods is empty; nothing to dump. < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:50:13.596 (6.13s) > Enter [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/12/25 14:50:13.596 STEP: Delete manifests - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:50 @ 10/12/25 14:50:13.596 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 14:50:13.596 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 14:50:15.993 (2.397s) < Exit [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/12/25 14:50:15.993 (2.397s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/12/25 14:50:15.994 + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/12/25 14:50:17.927 + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/12/25 14:50:15.994 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/12/25 14:50:17.927 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/12/25 14:50:17.927 (1.933s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/12/25 14:50:17.927 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/12/25 14:50:24.014 (6.087s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:100 @ 10/12/25 14:50:24.014 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:111 @ 10/12/25 14:50:24.014 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:122 @ 10/12/25 14:50:24.014 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:133 @ 10/12/25 14:50:24.014 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:144 @ 10/12/25 14:50:24.014 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:164 @ 10/12/25 14:50:24.014 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:185 @ 10/12/25 14:50:24.015 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:229 @ 10/12/25 14:50:24.015 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:290 @ 10/12/25 14:50:24.015 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:341 @ 10/12/25 14:50:24.015 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:367 @ 10/12/25 14:50:24.015 + + + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/12/25 14:50:24.015 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/12/25 14:50:25.485 (1.47s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:62 @ 10/12/25 14:50:25.485 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:62 @ 10/12/25 14:50:53.854 (28.369s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 14:50:53.854 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 14:50:53.854 (0s) + + + > Enter [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/12/25 14:50:53.854 < Exit [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/12/25 14:50:55.095 (1.24s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 14:50:55.095 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 14:50:55.095 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/12/25 14:50:55.095 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:102 @ 10/12/25 14:50:55.095 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/12/25 14:51:06.641 (11.546s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 14:51:06.641 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 14:51:06.641 (0s) + + + > Enter [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/12/25 14:51:06.641 STEP: CVIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:113 @ 10/12/25 14:51:06.641 < Exit [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/12/25 14:51:09.41 (2.769s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 14:51:09.41 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 14:51:09.411 (0s) + + + > Enter [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/12/25 14:51:09.411 STEP: VMClasses should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:124 @ 10/12/25 14:51:09.411 < Exit [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/12/25 14:51:12.158 (2.748s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 14:51:12.158 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 14:51:12.158 (0s) + + + [FAILED] Timed out after 344.487s. Expected success, but got an error: <*errors.errorString | 0x14000cbe900>: cannot patch VMIP "head-5073ae15-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip { s: "cannot patch VMIP \"head-5073ae15-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:138 @ 10/12/25 14:56:56.687 + > Enter [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/12/25 14:51:12.159 [FAILED] Timed out after 344.487s. Expected success, but got an error: <*errors.errorString | 0x14000cbe900>: cannot patch VMIP "head-5073ae15-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip { s: "cannot patch VMIP \"head-5073ae15-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:138 @ 10/12/25 14:56:56.687 < Exit [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/12/25 14:56:56.689 (5m44.489s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 14:56:56.689 The list of pods is empty; nothing to dump. < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 14:57:02.674 (5.985s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:141 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:152 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:162 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:173 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:184 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:196 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:220 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:245 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:268 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:280 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:299 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:316 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:341 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:354 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:358 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:375 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:388 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:405 @ 10/12/25 14:57:02.676 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:444 @ 10/12/25 14:57:02.677 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:479 @ 10/12/25 14:57:02.677 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:497 @ 10/12/25 14:57:02.677 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:512 @ 10/12/25 14:57:02.677 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:528 @ 10/12/25 14:57:02.677 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:545 @ 10/12/25 14:57:02.677 + + + > Enter [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/12/25 14:57:02.677 < Exit [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/12/25 14:57:04.163 (1.486s) > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/12/25 14:57:04.163 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/12/25 14:57:04.163 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/12/25 14:57:04.163 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/12/25 14:57:08.015 (3.853s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/12/25 14:57:08.015 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/12/25 14:57:08.015 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-safe --namespace head-5073ae15-end-to-end-vm-restore-safe --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-safe\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:107 @ 10/12/25 15:13:50.21 + > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/12/25 14:57:08.016 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/12/25 14:57:08.016 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/12/25 14:57:08.016 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/12/25 14:57:08.016 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/12/25 15:13:50.21 (16m42.195s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-safe --namespace head-5073ae15-end-to-end-vm-restore-safe --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-safe\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:107 @ 10/12/25 15:13:50.21 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/12/25 15:13:50.211 (16m42.196s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/12/25 15:13:50.211 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/12/25 15:13:56.356 (6.144s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:115 @ 10/12/25 15:13:56.356 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:158 @ 10/12/25 15:13:56.356 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:348 @ 10/12/25 15:13:56.356 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:374 @ 10/12/25 15:13:56.357 + + + > Enter [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/12/25 15:13:56.357 < Exit [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/12/25 15:13:59.512 (3.156s) > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/12/25 15:13:59.513 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/12/25 15:13:59.513 (0s) > Enter [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/12/25 15:13:59.513 STEP: Create a vmip automatically and check its binding with a lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:81 @ 10/12/25 15:13:59.513 STEP: Remove label from the lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:87 @ 10/12/25 15:14:00.421 STEP: Wait for the label to be restored by the controller - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:96 @ 10/12/25 15:14:00.547 < Exit [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/12/25 15:14:00.791 (1.278s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/12/25 15:14:00.791 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/12/25 15:14:00.791 (0s) + + + > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/12/25 15:14:00.791 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/12/25 15:14:00.791 (0s) > Enter [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/12/25 15:14:00.791 STEP: Create an intermediate vmip automatically to allocate a new ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:109 @ 10/12/25 15:14:00.791 STEP: Delete the intermediate vmip automatically and check that the lease is released - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:115 @ 10/12/25 15:14:01.16 STEP: Reuse the released lease with a static vmip - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:124 @ 10/12/25 15:14:01.406 STEP: Delete the static vmip and lease, then create another static vmip with this ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:132 @ 10/12/25 15:14:01.772 < Exit [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/12/25 15:14:02.55 (1.759s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/12/25 15:14:02.55 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/12/25 15:14:02.55 (0s) + + + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/12/25 15:14:02.55 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/12/25 15:14:04.553 (2.002s) > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 15:14:04.553 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 15:14:04.553 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/12/25 15:14:04.553 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/12/25 15:14:07.304 (2.751s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 15:14:07.304 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 15:14:07.304 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 15:14:07.305 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 15:14:07.305 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/12/25 15:14:07.305 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:75 @ 10/12/25 15:14:07.305 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/12/25 15:14:12.996 (5.691s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 15:14:12.996 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 15:14:12.996 (0s) + + + [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/12/25 15:14:51.702 This is the Progress Report generated when the suite timeout occurred: VirtualMachineLabelAndAnnotation When virtual disks are applied checks VDs phases (Spec Runtime: 38.705s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 In [It] (Node Runtime: 38.705s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 At [By Step] VDs should be in Ready phases (Step Runtime: 38.705s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 Spec Goroutine goroutine 894 [sync.WaitGroup.Wait] sync.runtime_SemacquireWaitGroup(0x100164b30?, 0x0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000b1c4c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140009a8260, 0x1, 0x101bdde58?}, {0x1018f6204, 0x28}, {{0x0, 0x0, 0x0}, 0x140005a93b0, {0x14000d82150, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1018f6204, 0x28}, {{0x0, 0x0, 0x0}, 0x140005a93b0, {0x14000d82150, 0x2c}, {0x14000ce7080, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1018f6204, 0x28}, {0x1018b4eef, 0x5}, {{0x0, 0x0, 0x0}, 0x140005a93b0, {0x14000d82150, 0x2c}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func18.6.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phases", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 37 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 980 [syscall] syscall.syscall6(0x100b1c500?, 0x12fa7bd48?, 0x1044d4f30?, 0x90?, 0x14000501008?, 0x14000b62360?, 0x14000bb4a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000bb4a88?, 0x1001e66fc?, 0x90?, 0x1021e2120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002d91f0?, 0x14000bb4ac4, 0x14000b1c500?, 0x140002d9180?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140000d2580) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000997408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000026600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000026600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400031b490?, 0x102279e88?, 0x140002d90a0?}}, {0x10229a488?, 0x140002d90a0?}, {0x1400098e410?, 0xc23305c491c25360?}, {0x102272d80, 0x140000d2500}, {0x102272d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400031b490?, 0x103787728?, 0x10375dfa0?}}, {0x10229a488, 0x140002d90a0}, {0x1400098e410, 0xcd}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10229ea38?, 0x14000521bd8?}, {0x1018b7fba?, 0x140000d23c0?}}, {0x1018f6204, 0x28}, {0x14000730a20, 0x26}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 894 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 15:14:12.997 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 15:14:12.997 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/12/25 15:14:12.997 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 @ 10/12/25 15:14:12.997 [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/12/25 15:14:51.702 This is the Progress Report generated when the suite timeout occurred: VirtualMachineLabelAndAnnotation When virtual disks are applied checks VDs phases (Spec Runtime: 38.705s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 In [It] (Node Runtime: 38.705s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 At [By Step] VDs should be in Ready phases (Step Runtime: 38.705s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 Spec Goroutine goroutine 894 [sync.WaitGroup.Wait] sync.runtime_SemacquireWaitGroup(0x100164b30?, 0x0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000b1c4c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140009a8260, 0x1, 0x101bdde58?}, {0x1018f6204, 0x28}, {{0x0, 0x0, 0x0}, 0x140005a93b0, {0x14000d82150, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1018f6204, 0x28}, {{0x0, 0x0, 0x0}, 0x140005a93b0, {0x14000d82150, 0x2c}, {0x14000ce7080, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1018f6204, 0x28}, {0x1018b4eef, 0x5}, {{0x0, 0x0, 0x0}, 0x140005a93b0, {0x14000d82150, 0x2c}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func18.6.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phases", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 37 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 980 [syscall] syscall.syscall6(0x100b1c500?, 0x12fa7bd48?, 0x1044d4f30?, 0x90?, 0x14000501008?, 0x14000b62360?, 0x14000bb4a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000bb4a88?, 0x1001e66fc?, 0x90?, 0x1021e2120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002d91f0?, 0x14000bb4ac4, 0x14000b1c500?, 0x140002d9180?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140000d2580) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000997408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000026600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000026600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400031b490?, 0x102279e88?, 0x140002d90a0?}}, {0x10229a488?, 0x140002d90a0?}, {0x1400098e410?, 0xc23305c491c25360?}, {0x102272d80, 0x140000d2500}, {0x102272d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400031b490?, 0x103787728?, 0x10375dfa0?}}, {0x10229a488, 0x140002d90a0}, {0x1400098e410, 0xcd}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10229ea38?, 0x14000521bd8?}, {0x1018b7fba?, 0x140000d23c0?}}, {0x1018f6204, 0x28}, {0x14000730a20, 0x26}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 894 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/12/25 15:14:51.706 (38.71s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 15:14:51.706 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 15:14:57.61 (5.903s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400000e4c8>: the container "virtualization-controller" was not found: virtualization-controller-66f75ff958-n6xkj the container "virtualization-controller" was restarted: virtualization-controller-8bdcbf866-n66wf the container "virtualization-controller" was restarted: virtualization-controller-8bdcbf866-px8hv { errs: [ <*errors.joinError | 0x1400000e498>{ errs: [ <*errors.joinError | 0x1400000e480>{ errs: [ <*errors.errorString | 0x14000f102c0>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-66f75ff958-n6xkj", }, ], }, <*errors.errorString | 0x14000f102f0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-8bdcbf866-n66wf", }, ], }, <*errors.errorString | 0x14000f10310>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-8bdcbf866-px8hv", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/12/25 15:14:57.898 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/12/25 15:14:57.618 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/12/25 15:14:57.618 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/12/25 15:14:57.618 [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400000e4c8>: the container "virtualization-controller" was not found: virtualization-controller-66f75ff958-n6xkj the container "virtualization-controller" was restarted: virtualization-controller-8bdcbf866-n66wf the container "virtualization-controller" was restarted: virtualization-controller-8bdcbf866-px8hv { errs: [ <*errors.joinError | 0x1400000e498>{ errs: [ <*errors.joinError | 0x1400000e480>{ errs: [ <*errors.errorString | 0x14000f102c0>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-66f75ff958-n6xkj", }, ], }, <*errors.errorString | 0x14000f102f0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-8bdcbf866-n66wf", }, ], }, <*errors.errorString | 0x14000f10310>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-8bdcbf866-px8hv", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/12/25 15:14:57.898 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/12/25 15:14:57.898 (280ms) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/12/25 15:14:57.898 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/12/25 15:15:18.979 (21.081s) + + + \ No newline at end of file diff --git a/artifacts/clean-cephrbd-20251013-081321-13799/junit.xml b/artifacts/clean-cephrbd-20251013-081321-13799/junit.xml new file mode 100644 index 0000000000..24fc1d6304 --- /dev/null +++ b/artifacts/clean-cephrbd-20251013-081321-13799/junit.xml @@ -0,0 +1,829 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 08:35:23.053 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 08:35:28.098 (5.044s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 08:35:28.098 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 08:35:28.098 (0s) + + + > Enter [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/13/25 08:35:28.1 < Exit [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/13/25 08:35:29.569 (1.47s) > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 08:35:29.569 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 08:35:29.569 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/13/25 08:35:29.569 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/13/25 08:35:36.486 (6.917s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 08:35:36.486 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 08:35:36.486 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 08:35:36.487 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 08:35:36.487 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/13/25 08:35:36.487 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:91 @ 10/13/25 08:35:36.487 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/13/25 08:35:42.178 (5.691s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 08:35:42.178 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 08:35:42.178 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 08:35:42.178 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 08:35:42.178 (0s) > Enter [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/13/25 08:35:42.178 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:102 @ 10/13/25 08:35:42.178 < Exit [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/13/25 08:36:30.601 (48.423s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 08:36:30.601 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 08:36:30.601 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 08:36:30.601 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 08:36:30.601 (0s) > Enter [It] checks VDs phases with map[vm:existing-vmclass] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/13/25 08:36:30.601 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:116 @ 10/13/25 08:36:30.602 < Exit [It] checks VDs phases with map[vm:existing-vmclass] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/13/25 08:36:33.374 (2.773s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 08:36:33.374 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 08:36:33.374 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 08:36:33.375 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 08:36:33.375 (0s) > Enter [It] checks VMs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:126 @ 10/13/25 08:36:33.375 STEP: VMs should be in Pending phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:127 @ 10/13/25 08:36:33.375 < Exit [It] checks VMs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:126 @ 10/13/25 08:36:38.884 (5.509s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 08:36:38.884 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 08:36:38.884 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-existing-vmclass --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-existing-vmclass\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:142 @ 10/13/25 08:53:21.025 + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 08:36:38.884 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 08:36:38.884 (0s) > Enter [It] checks VMs phases with map[vm:existing-vmclass] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:140 @ 10/13/25 08:36:38.884 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:141 @ 10/13/25 08:36:38.884 [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-existing-vmclass --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-existing-vmclass\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:142 @ 10/13/25 08:53:21.025 < Exit [It] checks VMs phases with map[vm:existing-vmclass] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:140 @ 10/13/25 08:53:21.026 (16m42.148s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 08:53:21.026 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 08:53:29.462 (8.436s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:152 @ 10/13/25 08:53:29.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:157 @ 10/13/25 08:53:29.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:163 @ 10/13/25 08:53:29.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:176 @ 10/13/25 08:53:29.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:181 @ 10/13/25 08:53:29.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:187 @ 10/13/25 08:53:29.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:202 @ 10/13/25 08:53:29.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:216 @ 10/13/25 08:53:29.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:240 @ 10/13/25 08:53:29.462 + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:51 @ 10/13/25 08:53:31.406 + > Enter [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/13/25 08:53:29.462 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:51 @ 10/13/25 08:53:31.406 < Exit [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/13/25 08:53:31.406 (1.944s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/13/25 08:53:31.406 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/13/25 08:53:37.049 (5.643s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:94 @ 10/13/25 08:53:37.049 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:103 @ 10/13/25 08:53:37.049 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:114 @ 10/13/25 08:53:37.049 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:128 @ 10/13/25 08:53:37.049 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:139 @ 10/13/25 08:53:37.049 + + + > Enter [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/13/25 08:53:37.05 < Exit [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/13/25 08:53:40.186 (3.136s) > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 08:53:40.186 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 08:53:40.186 (0s) > Enter [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/13/25 08:53:40.186 STEP: Create a vmip automatically and check its binding with a lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:81 @ 10/13/25 08:53:40.186 STEP: Remove label from the lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:87 @ 10/13/25 08:53:40.54 STEP: Wait for the label to be restored by the controller - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:96 @ 10/13/25 08:53:40.657 < Exit [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/13/25 08:53:40.885 (699ms) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 08:53:40.885 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 08:53:40.885 (0s) + + + > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 08:53:40.885 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 08:53:40.885 (0s) > Enter [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/13/25 08:53:40.885 STEP: Create an intermediate vmip automatically to allocate a new ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:109 @ 10/13/25 08:53:40.885 STEP: Delete the intermediate vmip automatically and check that the lease is released - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:115 @ 10/13/25 08:53:41.23 STEP: Reuse the released lease with a static vmip - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:124 @ 10/13/25 08:53:41.459 STEP: Delete the static vmip and lease, then create another static vmip with this ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:132 @ 10/13/25 08:53:41.802 < Exit [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/13/25 08:53:42.419 (1.534s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 08:53:42.419 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 08:53:42.419 (0s) + + + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/13/25 08:53:42.419 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/13/25 08:53:44.366 (1.948s) > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:53:44.367 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:53:44.367 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/13/25 08:53:44.367 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/13/25 08:53:47.565 (3.198s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:53:47.565 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:53:47.565 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:53:47.565 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:53:47.565 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/13/25 08:53:47.565 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:75 @ 10/13/25 08:53:47.565 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/13/25 08:53:53.272 (5.707s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:53:53.272 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:53:53.272 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:53:53.272 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:53:53.272 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/13/25 08:53:53.272 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 @ 10/13/25 08:53:53.272 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/13/25 08:54:37.488 (44.216s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:54:37.488 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:54:37.488 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:54:37.489 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:54:37.489 (0s) > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/13/25 08:54:37.489 STEP: Virtual machine phase should be Running - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:97 @ 10/13/25 08:54:37.489 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/13/25 08:54:44.397 (6.908s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:54:44.397 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:54:44.397 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:54:44.397 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:54:44.397 (0s) > Enter [It] marks VMs with label map["specialKey":"specialValue"] - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/13/25 08:54:44.397 < Exit [It] marks VMs with label map["specialKey":"specialValue"] - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/13/25 08:54:47.021 (2.624s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:54:47.021 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:54:47.022 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:54:47.022 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:54:47.022 (0s) > Enter [It] checks VMs and pods labels after VMs labeling - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/13/25 08:54:47.022 < Exit [It] checks VMs and pods labels after VMs labeling - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/13/25 08:54:49.279 (2.257s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:54:49.279 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:54:49.279 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:54:49.279 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:54:49.279 (0s) > Enter [It] removes label map[specialKey:specialValue] from VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/13/25 08:54:49.279 < Exit [It] removes label map[specialKey:specialValue] from VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/13/25 08:54:51.937 (2.657s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:54:51.937 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:54:51.937 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:54:51.937 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:54:51.937 (0s) > Enter [It] checks VMs and pods labels after VMs unlabeling - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/13/25 08:54:51.937 < Exit [It] checks VMs and pods labels after VMs unlabeling - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/13/25 08:54:54.164 (2.227s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:54:54.164 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:54:54.164 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:54:54.164 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:54:54.164 (0s) > Enter [It] marks VMs with annotation map["specialKey":"specialValue"] - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/13/25 08:54:54.164 < Exit [It] marks VMs with annotation map["specialKey":"specialValue"] - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/13/25 08:54:56.846 (2.681s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:54:56.846 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:54:56.846 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:54:56.846 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:54:56.846 (0s) > Enter [It] checks VMs and pods annotations after VMs annotating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/13/25 08:54:56.846 < Exit [It] checks VMs and pods annotations after VMs annotating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/13/25 08:54:59.109 (2.263s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:54:59.109 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:54:59.109 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:54:59.109 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:54:59.109 (0s) > Enter [It] removes annotation map[specialKey:specialValue] from VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/13/25 08:54:59.109 < Exit [It] removes annotation map[specialKey:specialValue] from VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/13/25 08:55:01.771 (2.662s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:55:01.771 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:55:01.771 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:55:01.772 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:55:01.772 (0s) > Enter [It] checks VMs and pods annotations after VMs unannotating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/13/25 08:55:01.772 < Exit [It] checks VMs and pods annotations after VMs unannotating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/13/25 08:55:04.017 (2.245s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:55:04.017 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:55:04.017 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:55:04.017 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 08:55:04.017 (0s) > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/13/25 08:55:04.017 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 08:55:04.017 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 08:55:52.858 (48.842s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/13/25 08:55:52.859 (48.842s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:55:52.859 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 08:55:52.859 (0s) + + + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/13/25 08:55:52.86 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/13/25 08:55:54.877 (2.017s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/13/25 08:55:54.877 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/13/25 08:55:58.783 (3.907s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:55:58.783 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:55:58.783 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/13/25 08:55:58.784 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:97 @ 10/13/25 08:55:58.784 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/13/25 08:56:12.647 (13.863s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:56:12.647 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:56:12.647 (0s) + + + > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/13/25 08:56:12.647 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 @ 10/13/25 08:56:12.647 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/13/25 08:57:15.663 (1m2.996s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:57:15.663 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:57:15.663 (0s) + + + > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:118 @ 10/13/25 08:57:15.663 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:119 @ 10/13/25 08:57:15.664 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:118 @ 10/13/25 08:58:34.476 (1m18.812s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:58:34.476 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:58:34.477 (0s) + + + > Enter [It] status should be in Running phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:129 @ 10/13/25 08:58:34.477 < Exit [It] status should be in Running phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:129 @ 10/13/25 08:58:40.021 (5.544s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:58:40.021 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:58:40.021 (0s) + + + > Enter [It] gets VMs and SVCs objects - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:142 @ 10/13/25 08:58:40.021 < Exit [It] gets VMs and SVCs objects - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:142 @ 10/13/25 08:58:44.499 (4.478s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:58:44.499 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:58:44.499 (0s) + + + > Enter [It] check ssh connection via `d8 v` to VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:166 @ 10/13/25 08:58:44.499 STEP: VirtualMachine "head-5073ae15-vm-connectivity-a" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:169 @ 10/13/25 08:58:44.499 STEP: VirtualMachine "head-5073ae15-vm-connectivity-b" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:169 @ 10/13/25 08:58:47.062 < Exit [It] check ssh connection via `d8 v` to VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:166 @ 10/13/25 08:58:49.888 (5.389s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:58:49.888 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:58:49.888 (0s) + + + > Enter [It] checks VMs connection to external network - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:174 @ 10/13/25 08:58:49.889 STEP: Cilium agent should be OK's for VM: head-5073ae15-vm-connectivity-a - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:175 @ 10/13/25 08:58:49.889 STEP: Cilium agent should be OK's for VM: head-5073ae15-vm-connectivity-b - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:175 @ 10/13/25 08:59:00.195 STEP: Response code from "https://flant.com" should be "200" for "head-5073ae15-vm-connectivity-a" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:176 @ 10/13/25 08:59:10.095 STEP: Response code from "https://flant.com" should be "200" for "head-5073ae15-vm-connectivity-b" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:176 @ 10/13/25 08:59:13.02 < Exit [It] checks VMs connection to external network - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:174 @ 10/13/25 08:59:14.854 (24.966s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:14.854 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:14.854 (0s) + + + > Enter [It] check nginx status via `d8 v` on VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:179 @ 10/13/25 08:59:14.855 STEP: VirtualMachine "head-5073ae15-vm-connectivity-a" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:182 @ 10/13/25 08:59:14.855 STEP: VirtualMachine "head-5073ae15-vm-connectivity-b" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:182 @ 10/13/25 08:59:16.773 < Exit [It] check nginx status via `d8 v` on VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:179 @ 10/13/25 08:59:18.377 (3.522s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:18.377 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:18.377 (0s) + + + > Enter [It] gets page from service head-5073ae15-vm-connectivity-a - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:187 @ 10/13/25 08:59:18.377 < Exit [It] gets page from service head-5073ae15-vm-connectivity-a - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:187 @ 10/13/25 08:59:19.702 (1.324s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:19.702 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:19.702 (0s) + + + > Enter [It] gets page from service head-5073ae15-vm-connectivity-b - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:198 @ 10/13/25 08:59:19.702 < Exit [It] gets page from service head-5073ae15-vm-connectivity-b - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:198 @ 10/13/25 08:59:21.556 (1.854s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:21.556 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:21.556 (0s) + + + > Enter [It] changes selector in service head-5073ae15-vm-connectivity-a with selector from service head-5073ae15-vm-connectivity-b - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:209 @ 10/13/25 08:59:21.624 < Exit [It] changes selector in service head-5073ae15-vm-connectivity-a with selector from service head-5073ae15-vm-connectivity-b - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:209 @ 10/13/25 08:59:22.522 (897ms) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:22.522 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:22.522 (0s) + + + > Enter [It] checks selector in service head-5073ae15-vm-connectivity-a - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:222 @ 10/13/25 08:59:22.522 STEP: Selector should be "vm-connectivity-b" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:223 @ 10/13/25 08:59:22.522 < Exit [It] checks selector in service head-5073ae15-vm-connectivity-a - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:222 @ 10/13/25 08:59:23.514 (993ms) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:23.515 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:23.515 (0s) + + + > Enter [It] gets page from service head-5073ae15-vm-connectivity-a - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:228 @ 10/13/25 08:59:23.515 STEP: Response should be from virtual machine "head-5073ae15-vm-connectivity-b" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:229 @ 10/13/25 08:59:23.515 < Exit [It] gets page from service head-5073ae15-vm-connectivity-a - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:228 @ 10/13/25 08:59:25.493 (1.978s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:25.493 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:25.493 (0s) + + + > Enter [It] changes back selector in service head-5073ae15-vm-connectivity-a - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:240 @ 10/13/25 08:59:25.494 < Exit [It] changes back selector in service head-5073ae15-vm-connectivity-a - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:240 @ 10/13/25 08:59:26.733 (1.24s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:26.733 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:26.734 (0s) + + + > Enter [It] checks selector in service head-5073ae15-vm-connectivity-a - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:250 @ 10/13/25 08:59:26.734 STEP: Selector should be "vm-connectivity-a" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:251 @ 10/13/25 08:59:26.734 < Exit [It] checks selector in service head-5073ae15-vm-connectivity-a - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:250 @ 10/13/25 08:59:27.694 (960ms) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:27.694 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 08:59:27.694 (0s) + + + > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:258 @ 10/13/25 08:59:27.695 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 08:59:27.695 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 09:00:14.294 (46.599s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:258 @ 10/13/25 09:00:14.294 (46.599s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 09:00:14.294 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 09:00:14.294 (0s) + + + > Enter [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/13/25 09:00:14.294 < Exit [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/13/25 09:00:14.295 (1ms) > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:00:14.295 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:00:14.295 (0s) > Enter [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/13/25 09:00:14.295 < Exit [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/13/25 09:00:15.56 (1.265s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:00:15.56 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:00:15.56 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:00:15.561 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:00:15.561 (0s) > Enter [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/13/25 09:00:15.561 STEP: Project should be deployed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:73 @ 10/13/25 09:00:15.561 < Exit [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/13/25 09:00:18.277 (2.716s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:00:18.277 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:00:18.278 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:00:18.278 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:00:18.278 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/13/25 09:00:18.278 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/13/25 09:00:21.406 (3.128s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:00:21.406 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:00:21.406 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:00:21.406 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:00:21.406 (0s) > Enter [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/13/25 09:00:21.406 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/13/25 09:00:21.407 < Exit [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/13/25 09:00:38.685 (17.278s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:00:38.685 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:00:38.685 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:00:38.685 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:00:38.685 (0s) > Enter [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/13/25 09:00:38.685 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/13/25 09:00:38.685 < Exit [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/13/25 09:01:54.409 (1m15.724s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:01:54.409 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:01:54.409 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:01:54.409 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:01:54.409 (0s) > Enter [It] When virtual machines are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/13/25 09:01:54.409 STEP: VMs should be in Running phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/13/25 09:01:54.409 < Exit [It] When virtual machines are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/13/25 09:02:12.583 (18.174s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:02:12.583 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:02:12.583 (0s) > Enter [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/13/25 09:02:12.583 STEP: Delete manifests - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:50 @ 10/13/25 09:02:12.584 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 09:02:12.584 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 09:03:24.104 (1m11.52s) < Exit [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/13/25 09:03:24.104 (1m11.52s) + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/13/25 09:03:26.052 + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/13/25 09:03:24.104 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/13/25 09:03:26.052 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/13/25 09:03:26.053 (1.948s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/13/25 09:03:26.053 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/13/25 09:03:31.771 (5.719s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:100 @ 10/13/25 09:03:31.772 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:111 @ 10/13/25 09:03:31.772 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:122 @ 10/13/25 09:03:31.772 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:133 @ 10/13/25 09:03:31.772 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:144 @ 10/13/25 09:03:31.772 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:164 @ 10/13/25 09:03:31.772 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:185 @ 10/13/25 09:03:31.772 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:229 @ 10/13/25 09:03:31.772 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:290 @ 10/13/25 09:03:31.772 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:341 @ 10/13/25 09:03:31.772 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:367 @ 10/13/25 09:03:31.772 + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.772 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 09:03:31.772 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.772 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.772 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 09:03:31.772 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.772 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.772 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 09:03:31.772 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.772 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.772 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 09:03:31.772 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.772 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.772 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 09:03:31.772 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.772 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.772 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 09:03:31.772 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.772 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.772 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 09:03:31.773 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.773 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.773 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 09:03:31.773 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.773 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.773 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 09:03:31.773 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 09:03:31.773 (0s) + + + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/13/25 09:03:31.773 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/13/25 09:03:33.742 (1.969s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:62 @ 10/13/25 09:03:33.742 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:62 @ 10/13/25 09:04:01.507 (27.765s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 09:04:01.507 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 09:04:01.507 (0s) + + + > Enter [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/13/25 09:04:01.508 < Exit [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/13/25 09:04:03.04 (1.532s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 09:04:03.04 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 09:04:03.04 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/13/25 09:04:03.04 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:102 @ 10/13/25 09:04:03.04 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/13/25 09:04:09.119 (6.079s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 09:04:09.119 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 09:04:09.119 (0s) + + + > Enter [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/13/25 09:04:09.119 STEP: CVIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:113 @ 10/13/25 09:04:09.119 < Exit [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/13/25 09:04:12.021 (2.902s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 09:04:12.021 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 09:04:12.021 (0s) + + + > Enter [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/13/25 09:04:12.021 STEP: VMClasses should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:124 @ 10/13/25 09:04:12.021 < Exit [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/13/25 09:04:14.698 (2.676s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 09:04:14.698 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 09:04:14.698 (0s) + + + [FAILED] Timed out after 344.759s. Expected success, but got an error: <*errors.errorString | 0x140006b60c0>: cannot patch VMIP "head-5073ae15-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip { s: "cannot patch VMIP \"head-5073ae15-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:138 @ 10/13/25 09:09:59.457 + > Enter [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/13/25 09:04:14.698 [FAILED] Timed out after 344.759s. Expected success, but got an error: <*errors.errorString | 0x140006b60c0>: cannot patch VMIP "head-5073ae15-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip { s: "cannot patch VMIP \"head-5073ae15-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:138 @ 10/13/25 09:09:59.457 < Exit [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/13/25 09:09:59.458 (5m44.761s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 09:09:59.458 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 09:10:33.64 (34.182s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:141 @ 10/13/25 09:10:33.64 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:152 @ 10/13/25 09:10:33.64 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:162 @ 10/13/25 09:10:33.641 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:173 @ 10/13/25 09:10:33.641 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:184 @ 10/13/25 09:10:33.641 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:196 @ 10/13/25 09:10:33.641 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:220 @ 10/13/25 09:10:33.641 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:245 @ 10/13/25 09:10:33.641 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:268 @ 10/13/25 09:10:33.641 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:280 @ 10/13/25 09:10:33.641 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:299 @ 10/13/25 09:10:33.641 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:316 @ 10/13/25 09:10:33.641 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:341 @ 10/13/25 09:10:33.641 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:354 @ 10/13/25 09:10:33.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:358 @ 10/13/25 09:10:33.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:375 @ 10/13/25 09:10:33.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:388 @ 10/13/25 09:10:33.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:405 @ 10/13/25 09:10:33.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:444 @ 10/13/25 09:10:33.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:479 @ 10/13/25 09:10:33.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:497 @ 10/13/25 09:10:33.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:512 @ 10/13/25 09:10:33.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:528 @ 10/13/25 09:10:33.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:545 @ 10/13/25 09:10:33.642 + + + > Enter [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/13/25 09:10:33.642 < Exit [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/13/25 09:10:35.609 (1.967s) > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 09:10:35.609 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 09:10:35.609 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/13/25 09:10:35.609 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/13/25 09:10:41.739 (6.13s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 09:10:41.739 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 09:10:41.74 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Attached'' state before 5m0s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait VirtualMachineBlockDeviceAttachment head-5073ae15-blank-disk-attachment-always-on-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Attached' --timeout=5m0s\nstderr: error: timed out waiting for the condition on virtualmachineblockdeviceattachments/head-5073ae15-blank-disk-attachment-always-on-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Attached'", "cmd: kubectl wait VirtualMachineBlockDeviceAttachment head-5073ae15-blank-disk-attachment-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Attached' --timeout=5m0s\nstderr: error: timed out waiting for the condition on virtualmachineblockdeviceattachments/head-5073ae15-blank-disk-attachment-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Attached'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:120 @ 10/13/25 09:17:38.281 + > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 09:10:41.74 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 09:10:41.74 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/13/25 09:10:41.74 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/13/25 09:10:41.74 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/13/25 09:12:36.345 (1m54.581s) STEP: `VirtualMachineBlockDeviceAttachment` should be attached - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:119 @ 10/13/25 09:12:36.345 END STEP: `VirtualMachineBlockDeviceAttachment` should be attached - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:119 @ 10/13/25 09:17:38.28 (5m1.914s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Attached'' state before 5m0s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait VirtualMachineBlockDeviceAttachment head-5073ae15-blank-disk-attachment-always-on-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Attached' --timeout=5m0s\nstderr: error: timed out waiting for the condition on virtualmachineblockdeviceattachments/head-5073ae15-blank-disk-attachment-always-on-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Attached'", "cmd: kubectl wait VirtualMachineBlockDeviceAttachment head-5073ae15-blank-disk-attachment-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Attached' --timeout=5m0s\nstderr: error: timed out waiting for the condition on virtualmachineblockdeviceattachments/head-5073ae15-blank-disk-attachment-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Attached'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:120 @ 10/13/25 09:17:38.281 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/13/25 09:17:38.283 (6m56.497s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 09:17:38.283 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-restore-force -o yaml -l 'testcase=vm-restore-force' --show-managed-fields=true error: signal: killed stderr: E1013 09:17:48.336979 12769 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"https://api.dvp-e2e-clean-cephrbd-20251013-081321-13799.clean-cephrbd-20251013-081321-13799.e2e.virtlab.flant.com/api?timeout=32s\": net/http: TLS handshake timeout" E1013 09:17:58.343205 12769 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"https://api.dvp-e2e-clean-cephrbd-20251013-081321-13799.clean-cephrbd-20251013-081321-13799.e2e.virtlab.flant.com/api?timeout=32s\": net/http: TLS handshake timeout" Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-restore-force -o json -l 'testcase=vm-restore-force' stderr: E1013 09:18:18.340800 12792 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"https://api.dvp-e2e-clean-cephrbd-20251013-081321-13799.clean-cephrbd-20251013-081321-13799.e2e.virtlab.flant.com/api?timeout=32s\": net/http: TLS handshake timeout" The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 09:18:38.3 (1m0.017s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:131 @ 10/13/25 09:18:38.301 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:177 @ 10/13/25 09:18:38.301 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:341 @ 10/13/25 09:18:38.301 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:367 @ 10/13/25 09:18:38.301 + + + + > Enter [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/13/25 09:18:38.301 [SKIPPED] Module SDN is disabled. Skipping all tests for module SDN. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:50 @ 10/13/25 09:19:08.302 < Exit [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/13/25 09:19:08.304 (30.003s) > Enter [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/13/25 09:19:08.304 < Exit [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/13/25 09:19:08.304 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/13/25 09:19:08.305 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/13/25 09:19:08.305 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:111 @ 10/13/25 09:19:08.305 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:125 @ 10/13/25 09:19:08.305 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:141 @ 10/13/25 09:19:08.305 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:156 @ 10/13/25 09:19:08.305 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:169 @ 10/13/25 09:19:08.305 + + + + > Enter [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/13/25 09:19:08.305 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:41 @ 10/13/25 09:19:08.306 < Exit [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/13/25 09:19:08.306 (0s) > Enter [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/13/25 09:19:08.306 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 09:19:08.306 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 09:22:02.839 (2m54.532s) [FAILED] cannot delete test case resources kustomizationDir: /tmp/testdata/vm-migration-cancel cmd: kubectl delete --kustomize /tmp/testdata/vm-migration-cancel --ignore-not-found stderr: E1013 09:19:38.392445 12845 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"https://api.dvp-e2e-clean-cephrbd-20251013-081321-13799.clean-cephrbd-20251013-081321-13799.e2e.virtlab.flant.com/api?timeout=32s\": dial tcp: lookup api.dvp-e2e-clean-cephrbd-20251013-081321-13799.clean-cephrbd-20251013-081321-13799.e2e.virtlab.flant.com: i/o timeout" E1013 09:20:08.395337 12845 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"https://api.dvp-e2e-clean-cephrbd-20251013-081321-13799.clean-cephrbd-20251013-081321-13799.e2e.virtlab.flant.com/api?timeout=32s\": dial tcp: lookup api.dvp-e2e-clean-cephrbd-20251013-081321-13799.clean-cephrbd-20251013-081321-13799.e2e.virtlab.flant.com: i/o timeout" error when deleting "/tmp/testdata/vm-migration-cancel": Delete "https://api.dvp-e2e-clean-cephrbd-20251013-081321-13799.clean-cephrbd-20251013-081321-13799.e2e.virtlab.flant.com/api/v1/namespaces/head-5073ae15-end-to-end-vm-migration-cancel/secrets/head-5073ae15-cloud-init-migration-bios": dial tcp: lookup api.dvp-e2e-clean-cephrbd-20251013-081321-13799.clean-cephrbd-20251013-081321-13799.e2e.virtlab.flant.com: i/o timeout error when deleting "/tmp/testdata/vm-migration-cancel": Delete "https://api.dvp-e2e-clean-cephrbd-20251013-081321-13799.clean-cephrbd-20251013-081321-13799.e2e.virtlab.flant.com/api/v1/namespaces/head-5073ae15-end-to-end-vm-migration-cancel/secrets/head-5073ae15-cloud-init-migration-uefi": dial tcp: lookup api.dvp-e2e-clean-cephrbd-20251013-081321-13799.clean-cephrbd-20251013-081321-13799.e2e.virtlab.flant.com: i/o timeout error when deleting "/tmp/testdata/vm-migration-cancel": Delete "https://api.dvp-e2e-clean-cephrbd-20251013-081321-13799.clean-cephrbd-20251013-081321-13799.e2e.virtlab.flant.com/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-migration-cancel/virtualdisks/head-5073ae15-vd-blank-migration-bios": dial tcp: lookup api.dvp-e2e-clean-cephrbd-20251013-081321-13799.clean-cephrbd-20251013-081321-13799.e2e.virtlab.flant.com: i/o timeout error when deleting "/tmp/testdata/vm-migration-cancel": Delete "https://api.dvp-e2e-clean-cephrbd-20251013-081321-13799.clean-cephrbd-20251013-081321-13799.e2e.virtlab.flant.com/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-migration-cancel/virtualdisks/head-5073ae15-vd-blank-migration-uefi": net/http: TLS handshake timeout Unexpected error: <*exec.ExitError | 0x14000a10000>: exit status 1 { ProcessState: { pid: 12845, status: 256, rusage: { Utime: { Sec: 0, Usec: 137786, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 56170, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 84508672, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5958, Majflt: 19, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 60, Msgrcv: 60, Nsignals: 140, Nvcsw: 237, Nivcsw: 2092, }, }, Stderr: nil, } occurred In [AfterEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:617 @ 10/13/25 09:22:02.839 < Exit [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/13/25 09:22:02.839 (2m54.533s) + + + > Enter [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/13/25 09:22:02.839 < Exit [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/13/25 09:22:05.312 (2.473s) > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/13/25 09:22:05.312 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/13/25 09:22:05.312 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/13/25 09:22:05.312 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/13/25 09:22:09.784 (4.471s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 09:22:09.784 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 09:22:09.784 (0s) + + + > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/13/25 09:22:09.786 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/13/25 09:22:09.786 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/13/25 09:22:09.786 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/13/25 09:22:09.786 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/13/25 09:24:15.362 (2m5.575s) < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/13/25 09:24:15.362 (2m5.576s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 09:24:15.362 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 09:24:15.362 (0s) + + + + > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/13/25 09:24:15.364 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/13/25 09:24:15.364 (0s) > Enter [It] add additional interface to virtual machines - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:115 @ 10/13/25 09:24:15.364 [SKIPPED] Module SDN is disabled. Skipping part of tests. In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:118 @ 10/13/25 09:24:15.974 < Exit [It] add additional interface to virtual machines - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:115 @ 10/13/25 09:24:15.974 (610ms) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 09:24:15.974 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 09:24:15.974 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 5m0s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachinesnapshots head-5073ae15-vm-restore-safe --namespace head-5073ae15-end-to-end-vm-restore-safe --for='jsonpath={.status.phase}=Ready' --timeout=5m0s\nstderr: error: timed out waiting for the condition on virtualmachinesnapshots/head-5073ae15-vm-restore-safe\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:180 @ 10/13/25 09:29:19.072 + > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/13/25 09:24:15.975 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/13/25 09:24:15.975 (0s) > Enter [It] restore the `VirtualMachines` with `Safe` mode - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:158 @ 10/13/25 09:24:15.975 STEP: Getting `VirtualMachines` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:162 @ 10/13/25 09:24:15.975 END STEP: Getting `VirtualMachines` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:162 @ 10/13/25 09:24:17.096 (1.121s) STEP: Creating `VirtualMachineSnapshots` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:170 @ 10/13/25 09:24:17.096 END STEP: Creating `VirtualMachineSnapshots` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:170 @ 10/13/25 09:29:19.072 (5m2.004s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 5m0s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachinesnapshots head-5073ae15-vm-restore-safe --namespace head-5073ae15-end-to-end-vm-restore-safe --for='jsonpath={.status.phase}=Ready' --timeout=5m0s\nstderr: error: timed out waiting for the condition on virtualmachinesnapshots/head-5073ae15-vm-restore-safe\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:180 @ 10/13/25 09:29:19.072 < Exit [It] restore the `VirtualMachines` with `Safe` mode - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:158 @ 10/13/25 09:29:19.073 (5m3.127s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 09:29:19.073 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 09:29:28.62 (9.547s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:348 @ 10/13/25 09:29:28.621 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:374 @ 10/13/25 09:29:28.621 + + + + > Enter [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/13/25 09:29:28.621 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:51 @ 10/13/25 09:29:28.621 < Exit [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/13/25 09:29:28.621 (0s) > Enter [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/13/25 09:29:28.621 < Exit [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/13/25 09:29:28.621 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:82 @ 10/13/25 09:29:28.621 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:94 @ 10/13/25 09:29:28.622 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:106 @ 10/13/25 09:29:28.622 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:116 @ 10/13/25 09:29:28.622 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:132 @ 10/13/25 09:29:28.622 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:151 @ 10/13/25 09:29:28.622 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:157 @ 10/13/25 09:29:28.622 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:189 @ 10/13/25 09:29:28.622 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:216 @ 10/13/25 09:29:28.622 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:252 @ 10/13/25 09:29:28.622 + + + > Enter [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/13/25 09:29:28.622 < Exit [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/13/25 09:29:30.585 (1.963s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/13/25 09:29:30.586 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/13/25 09:29:36.74 (6.155s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 09:29:36.74 < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 09:29:36.74 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:77 @ 10/13/25 09:46:18.877 + > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/13/25 09:29:36.742 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:76 @ 10/13/25 09:29:36.742 [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:77 @ 10/13/25 09:46:18.877 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/13/25 09:46:18.878 (16m42.101s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 09:46:18.878 < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 09:46:30.582 (11.704s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:86 @ 10/13/25 09:46:30.582 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:100 @ 10/13/25 09:46:30.582 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:116 @ 10/13/25 09:46:30.582 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:131 @ 10/13/25 09:46:30.582 + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 @ 10/13/25 10:03:20.086 + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/13/25 09:46:30.582 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/13/25 09:46:32.599 (2.017s) > Enter [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/13/25 09:46:32.599 < Exit [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/13/25 09:46:38.002 (5.403s) > Enter [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/13/25 09:46:38.002 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 @ 10/13/25 09:46:38.002 [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 @ 10/13/25 10:03:20.086 < Exit [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/13/25 10:03:20.086 (16m42.079s) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/13/25 10:03:20.086 The list of pods is empty; nothing to dump. STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 10:03:26.494 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 10:03:31.211 (4.718s) < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/13/25 10:03:31.212 (11.125s) + + + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/13/25 10:03:31.212 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/13/25 10:03:33.212 (2s) > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 10:03:33.213 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 10:03:33.213 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/13/25 10:03:33.213 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/13/25 10:03:36.568 (3.356s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 10:03:36.568 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 10:03:36.568 (0s) + + + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 10:03:36.569 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 10:03:36.569 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/13/25 10:03:36.569 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:88 @ 10/13/25 10:03:36.569 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/13/25 10:03:42.359 (5.789s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 10:03:42.359 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 10:03:42.359 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-with-hotplug-standalone --namespace head-5073ae15-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/13/25 10:20:24.434 + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 10:03:42.359 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 10:03:42.359 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/13/25 10:03:42.359 STEP: VDs with consumers should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:99 @ 10/13/25 10:03:42.359 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-with-hotplug-standalone --namespace head-5073ae15-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/13/25 10:20:24.434 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/13/25 10:20:24.434 (16m42.069s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 10:20:24.434 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 10:20:31.323 (6.89s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/13/25 10:20:31.324 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/13/25 10:20:31.324 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/13/25 10:20:31.324 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/13/25 10:20:31.324 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/13/25 10:20:31.324 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/13/25 10:20:31.324 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/13/25 10:20:31.324 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/13/25 10:20:31.325 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/13/25 10:20:31.325 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/13/25 10:20:31.325 + + + > Enter [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/13/25 10:20:31.325 < Exit [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/13/25 10:20:33.401 (2.076s) > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 10:20:33.401 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 10:20:33.401 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/13/25 10:20:33.401 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/13/25 10:20:35.788 (2.387s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 10:20:35.788 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 10:20:35.788 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root --namespace head-5073ae15-end-to-end-vm-versions --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1013 10:37:08.114835 17189 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: stream error: stream ID 1; INTERNAL_ERROR; received from peer\") has prevented the request from succeeding\nW1013 10:37:10.267528 17189 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding\nE1013 10:37:10.267618 17189 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: an error on the server (\\\"<html>\\\\r\\\\n<head><title>502 Bad Gateway</title></head>\\\\r\\\\n<body>\\\\r\\\\n<center><h1>502 Bad Gateway</h1></center>\\\\r\\\\n<hr><center>nginx</center>\\\\r\\\\n</body>\\\\r\\\\n</html>\\\") has prevented the request from succeeding\"\nW1013 10:37:12.069036 17189 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding\nE1013 10:37:12.069170 17189 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: an error on the server (\\\"<html>\\\\r\\\\n<head><title>502 Bad Gateway</title></head>\\\\r\\\\n<body>\\\\r\\\\n<center><h1>502 Bad Gateway</h1></center>\\\\r\\\\n<hr><center>nginx</center>\\\\r\\\\n</body>\\\\r\\\\n</html>\\\") has prevented the request from succeeding\"\nerror: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:69 @ 10/13/25 10:37:17.832 + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 10:20:35.788 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 10:20:35.788 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/13/25 10:20:35.788 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:68 @ 10/13/25 10:20:35.788 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root --namespace head-5073ae15-end-to-end-vm-versions --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1013 10:37:08.114835 17189 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: stream error: stream ID 1; INTERNAL_ERROR; received from peer\") has prevented the request from succeeding\nW1013 10:37:10.267528 17189 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding\nE1013 10:37:10.267618 17189 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: an error on the server (\\\"<html>\\\\r\\\\n<head><title>502 Bad Gateway</title></head>\\\\r\\\\n<body>\\\\r\\\\n<center><h1>502 Bad Gateway</h1></center>\\\\r\\\\n<hr><center>nginx</center>\\\\r\\\\n</body>\\\\r\\\\n</html>\\\") has prevented the request from succeeding\"\nW1013 10:37:12.069036 17189 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding\nE1013 10:37:12.069170 17189 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: an error on the server (\\\"<html>\\\\r\\\\n<head><title>502 Bad Gateway</title></head>\\\\r\\\\n<body>\\\\r\\\\n<center><h1>502 Bad Gateway</h1></center>\\\\r\\\\n<hr><center>nginx</center>\\\\r\\\\n</body>\\\\r\\\\n</html>\\\") has prevented the request from succeeding\"\nerror: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:69 @ 10/13/25 10:37:17.832 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/13/25 10:37:17.832 (16m42.095s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 10:37:17.832 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-versions -o yaml -l 'testcase=vm-versions' --show-managed-fields=true error: exit status 1 stderr: E1013 10:37:18.428331 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:18.565014 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:18.701543 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:18.836079 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:18.970867 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:18.972384 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:18.977806 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:18.979184 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:18.985152 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:18.986552 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:18.990568 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:18.991809 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:18.995706 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:18.996915 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:18.999536 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:19.136222 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:19.272016 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:19.405257 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:19.535962 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:19.667716 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:19.799572 18428 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-versions -o json -l 'testcase=vm-versions' stderr: E1013 10:37:20.483650 18434 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:20.623251 18434 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:20.778346 18434 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:20.929093 18434 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:20.933763 18434 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:20.935026 18434 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 10:37:21.089 (3.256s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/13/25 10:37:21.089 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/13/25 10:37:21.089 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/13/25 10:37:21.089 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/13/25 10:37:21.089 + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.089 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 10:37:21.089 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.089 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.089 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 10:37:21.089 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.089 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.09 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 10:37:21.09 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.09 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.09 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 10:37:21.09 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.09 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.09 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 10:37:21.09 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.09 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.09 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 10:37:21.09 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.09 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.09 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 10:37:21.09 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.09 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.09 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 10:37:21.09 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.09 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.09 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 10:37:21.09 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.09 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.09 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 10:37:21.091 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.091 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.091 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 10:37:21.091 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.091 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.091 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 10:37:21.091 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 10:37:21.091 (0s) + + + + > Enter [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/13/25 10:37:21.091 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:73 @ 10/13/25 10:37:21.091 < Exit [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/13/25 10:37:21.091 (0s) > Enter [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/13/25 10:37:21.091 < Exit [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/13/25 10:37:21.091 (0s) > Enter [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/13/25 10:37:21.091 < Exit [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/13/25 10:37:21.091 (0s) + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-affinity-toleration Unexpected error: <*exec.ExitError | 0x14000a104a0>: exit status 1 { ProcessState: { pid: 18435, status: 256, rusage: { Utime: { Sec: 0, Usec: 26097, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 10921, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46776320, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3660, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 11, Nsignals: 45, Nvcsw: 22, Nivcsw: 494, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:69 @ 10/13/25 10:37:21.688 + > Enter [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/13/25 10:37:21.091 [FAILED] kubectl create namespace head-5073ae15-end-to-end-affinity-toleration Unexpected error: <*exec.ExitError | 0x14000a104a0>: exit status 1 { ProcessState: { pid: 18435, status: 256, rusage: { Utime: { Sec: 0, Usec: 26097, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 10921, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46776320, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3660, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 11, Nsignals: 45, Nvcsw: 22, Nivcsw: 494, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:69 @ 10/13/25 10:37:21.688 < Exit [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/13/25 10:37:21.688 (597ms) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 10:37:21.688 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-affinity-toleration -o yaml -l 'testcase=affinity-toleration' --show-managed-fields=true error: exit status 1 stderr: E1013 10:37:22.622335 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:22.755050 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:22.889521 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.022285 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.156586 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.288961 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.421118 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.561863 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.696598 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.828029 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.961510 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.962962 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.967317 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.968623 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.971416 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.972734 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.975818 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.977028 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.978535 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.979840 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:23.981102 18438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-affinity-toleration -o json -l 'testcase=affinity-toleration' stderr: E1013 10:37:24.828564 18444 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:24.966422 18444 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:24.969221 18444 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:24.970655 18444 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:24.972894 18444 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:24.974075 18444 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 10:37:25.107 (3.419s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/13/25 10:37:25.107 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:120 @ 10/13/25 10:37:25.107 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:283 @ 10/13/25 10:37:25.107 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:372 @ 10/13/25 10:37:25.108 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:467 @ 10/13/25 10:37:25.108 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-configuration Unexpected error: <*exec.ExitError | 0x14000704080>: exit status 1 { ProcessState: { pid: 18450, status: 256, rusage: { Utime: { Sec: 0, Usec: 31027, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 17296, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46530560, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3646, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 41, Nvcsw: 21, Nivcsw: 586, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:55 @ 10/13/25 10:37:25.697 + > Enter [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/13/25 10:37:25.108 [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-configuration Unexpected error: <*exec.ExitError | 0x14000704080>: exit status 1 { ProcessState: { pid: 18450, status: 256, rusage: { Utime: { Sec: 0, Usec: 31027, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 17296, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46530560, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3646, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 41, Nvcsw: 21, Nivcsw: 586, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:55 @ 10/13/25 10:37:25.697 < Exit [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/13/25 10:37:25.697 (590ms) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 10:37:25.697 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-configuration -o yaml -l 'testcase=vm-configuration' --show-managed-fields=true error: exit status 1 stderr: E1013 10:37:26.629577 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.760643 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.893935 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.895376 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.898452 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.899936 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.903175 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.904511 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.908489 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.909743 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.912438 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.913847 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.917139 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.918351 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.921337 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.922590 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.926239 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.927500 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.928822 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.930086 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.931345 18451 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-configuration -o json -l 'testcase=vm-configuration' stderr: E1013 10:37:26.973202 18452 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.974485 18452 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.976196 18452 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.977351 18452 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.979437 18452 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:26.980654 18452 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 10:37:27.914 (2.216s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/13/25 10:37:27.914 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/13/25 10:37:27.914 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:109 @ 10/13/25 10:37:27.914 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:123 @ 10/13/25 10:37:27.914 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:147 @ 10/13/25 10:37:27.914 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:161 @ 10/13/25 10:37:27.914 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:183 @ 10/13/25 10:37:27.914 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:202 @ 10/13/25 10:37:27.914 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:226 @ 10/13/25 10:37:27.914 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:240 @ 10/13/25 10:37:27.914 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:250 @ 10/13/25 10:37:27.914 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:265 @ 10/13/25 10:37:27.914 + + + [FAILED] Unexpected error: <*exec.ExitError | 0x14000a110a0>: exit status 1 { ProcessState: { pid: 18456, status: 256, rusage: { Utime: { Sec: 0, Usec: 62314, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 53994, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 50905088, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3901, Majflt: 31, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 21, Msgrcv: 24, Nsignals: 133, Nvcsw: 151, Nivcsw: 1837, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:65 @ 10/13/25 10:37:29.678 + > Enter [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/13/25 10:37:27.914 [FAILED] Unexpected error: <*exec.ExitError | 0x14000a110a0>: exit status 1 { ProcessState: { pid: 18456, status: 256, rusage: { Utime: { Sec: 0, Usec: 62314, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 53994, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 50905088, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3901, Majflt: 31, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 21, Msgrcv: 24, Nsignals: 133, Nvcsw: 151, Nivcsw: 1837, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:65 @ 10/13/25 10:37:29.678 < Exit [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/13/25 10:37:29.678 (1.764s) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/13/25 10:37:29.678 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-image-hotplug -o yaml -l 'testcase=image-hotplug' --show-managed-fields=true error: exit status 1 stderr: E1013 10:37:30.242658 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.372625 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.509415 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.636469 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.765522 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.896871 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.899629 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.901588 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.904094 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.905635 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.909071 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.910459 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.913228 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.914427 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.916840 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.918341 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.921038 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.922242 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.923474 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.924808 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.926056 18457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-image-hotplug -o json -l 'testcase=image-hotplug' stderr: E1013 10:37:30.967377 18463 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.968710 18463 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.971233 18463 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.972489 18463 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.974645 18463 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 10:37:30.975753 18463 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/13/25 10:37:31.462 (1.784s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:112 @ 10/13/25 10:37:31.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:147 @ 10/13/25 10:37:31.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:190 @ 10/13/25 10:37:31.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:196 @ 10/13/25 10:37:31.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:204 @ 10/13/25 10:37:31.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:224 @ 10/13/25 10:37:31.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:236 @ 10/13/25 10:37:31.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:258 @ 10/13/25 10:37:31.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:284 @ 10/13/25 10:37:31.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:294 @ 10/13/25 10:37:31.462 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:309 @ 10/13/25 10:37:31.462 + + + [FAILED] Expected success, but got an error: <*errors.StatusError | 0x14000616460>: an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get pods) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding (get pods)", Reason: "InternalError", Details: { Name: "", Group: "", Kind: "pods", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 502, }, } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/13/25 10:37:31.965 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 10:37:31.462 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 10:37:31.462 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 10:37:31.462 [FAILED] Expected success, but got an error: <*errors.StatusError | 0x14000616460>: an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get pods) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding (get pods)", Reason: "InternalError", Details: { Name: "", Group: "", Kind: "pods", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 502, }, } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/13/25 10:37:31.965 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 10:37:31.965 (502ms) + + + [FAILED] Expected success, but got an error: <*errors.StatusError | 0x14000a5bd60>: an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get projects.deckhouse.io) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding (get projects.deckhouse.io)", Reason: "InternalError", Details: { Name: "", Group: "deckhouse.io", Kind: "projects", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 502, }, } In [DeferCleanup (Suite)] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:177 @ 10/13/25 10:37:32.108 + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/13/25 10:37:31.965 [FAILED] Expected success, but got an error: <*errors.StatusError | 0x14000a5bd60>: an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get projects.deckhouse.io) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding (get projects.deckhouse.io)", Reason: "InternalError", Details: { Name: "", Group: "deckhouse.io", Kind: "projects", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 502, }, } In [DeferCleanup (Suite)] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:177 @ 10/13/25 10:37:32.108 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/13/25 10:37:32.108 (143ms) + + + \ No newline at end of file diff --git a/artifacts/clean-cephrbd-20251013-173506-9941/junit.xml b/artifacts/clean-cephrbd-20251013-173506-9941/junit.xml new file mode 100644 index 0000000000..651d98c0a2 --- /dev/null +++ b/artifacts/clean-cephrbd-20251013-173506-9941/junit.xml @@ -0,0 +1,871 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 18:00:34.811 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 18:00:39.558 (4.75s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 18:00:39.558 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 18:00:39.558 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.56 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:00:39.56 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.56 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.56 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:00:39.56 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.56 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.56 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:00:39.56 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.56 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.56 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:00:39.56 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.56 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.56 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:00:39.56 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.56 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.56 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:00:39.56 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.561 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.561 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:00:39.561 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.561 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.561 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:00:39.561 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.561 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.561 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:00:39.561 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.561 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.561 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:00:39.561 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.561 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.561 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:00:39.561 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.561 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.561 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:00:39.561 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:00:39.561 (0s) + + + > Enter [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/13/25 18:00:39.561 < Exit [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/13/25 18:00:41.528 (1.968s) > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 18:00:41.528 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 18:00:41.528 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/13/25 18:00:41.528 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/13/25 18:00:48.508 (6.982s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 18:00:48.508 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 18:00:48.508 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-always-on-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-always-on-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 @ 10/13/25 18:17:30.566 + > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 18:00:48.508 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 18:00:48.509 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/13/25 18:00:48.509 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/13/25 18:00:48.509 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/13/25 18:17:30.566 (16m42.078s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-always-on-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-always-on-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 @ 10/13/25 18:17:30.566 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/13/25 18:17:30.566 (16m42.079s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 18:17:30.566 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 18:17:37.045 (6.479s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:131 @ 10/13/25 18:17:37.045 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:177 @ 10/13/25 18:17:37.045 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:341 @ 10/13/25 18:17:37.046 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:367 @ 10/13/25 18:17:37.046 + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.046 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 18:17:37.046 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.046 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.046 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 18:17:37.046 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.046 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.047 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 18:17:37.047 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.047 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.047 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 18:17:37.047 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.047 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.047 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 18:17:37.047 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.047 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.047 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 18:17:37.047 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.047 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.047 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 18:17:37.047 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.047 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.048 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 18:17:37.048 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.048 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.048 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 18:17:37.048 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 18:17:37.048 (0s) + + + > Enter [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/13/25 18:17:37.048 < Exit [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/13/25 18:17:39.112 (2.064s) > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 18:17:39.112 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 18:17:39.112 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/13/25 18:17:39.112 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/13/25 18:17:41.718 (2.607s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 18:17:41.718 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 18:17:41.719 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 18:17:41.719 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 18:17:41.719 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/13/25 18:17:41.719 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:68 @ 10/13/25 18:17:41.719 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/13/25 18:17:59.764 (18.046s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 18:17:59.764 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 18:17:59.765 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 18:17:59.765 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 18:17:59.765 (0s) > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/13/25 18:17:59.765 STEP: VM should be in Running phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:79 @ 10/13/25 18:17:59.765 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/13/25 18:18:02.621 (2.856s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 18:18:02.621 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 18:18:02.621 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 18:18:02.622 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 18:18:02.622 (0s) > Enter [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/13/25 18:18:02.622 < Exit [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/13/25 18:18:02.622 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 18:18:02.622 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 18:18:02.622 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 18:18:02.622 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 18:18:02.622 (0s) > Enter [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/13/25 18:18:02.622 < Exit [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/13/25 18:18:02.622 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 18:18:02.622 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 18:18:02.622 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 18:18:02.622 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 18:18:02.622 (0s) > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/13/25 18:18:02.622 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 18:18:02.622 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 18:19:13.05 (1m10.428s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/13/25 18:19:13.05 (1m10.428s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 18:19:13.05 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 18:19:13.05 (0s) + + + > Enter [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/13/25 18:19:13.05 < Exit [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/13/25 18:19:15.135 (2.085s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/13/25 18:19:15.135 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/13/25 18:19:20.35 (5.215s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 18:19:20.35 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 18:19:20.35 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/13/25 18:19:20.35 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:89 @ 10/13/25 18:19:20.35 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/13/25 18:19:25.343 (4.993s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 18:19:25.343 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 18:19:25.343 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:4>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-manual-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-manual-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:100 @ 10/13/25 18:36:07.383 + > Enter [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/13/25 18:19:25.343 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:4>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-manual-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-manual-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:100 @ 10/13/25 18:36:07.383 < Exit [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/13/25 18:36:07.383 (16m42.057s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 18:36:07.383 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 18:36:14.409 (7.026s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:109 @ 10/13/25 18:36:14.409 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:123 @ 10/13/25 18:36:14.41 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:147 @ 10/13/25 18:36:14.41 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:161 @ 10/13/25 18:36:14.41 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:183 @ 10/13/25 18:36:14.41 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:202 @ 10/13/25 18:36:14.41 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:226 @ 10/13/25 18:36:14.41 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:240 @ 10/13/25 18:36:14.41 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:250 @ 10/13/25 18:36:14.41 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:265 @ 10/13/25 18:36:14.411 + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/13/25 18:36:16.476 + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/13/25 18:36:14.411 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/13/25 18:36:16.476 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/13/25 18:36:16.476 (2.066s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/13/25 18:36:16.476 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/13/25 18:36:22.608 (6.132s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:100 @ 10/13/25 18:36:22.609 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:111 @ 10/13/25 18:36:22.609 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:122 @ 10/13/25 18:36:22.609 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:133 @ 10/13/25 18:36:22.609 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:144 @ 10/13/25 18:36:22.609 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:164 @ 10/13/25 18:36:22.609 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:185 @ 10/13/25 18:36:22.609 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:229 @ 10/13/25 18:36:22.609 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:290 @ 10/13/25 18:36:22.609 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:341 @ 10/13/25 18:36:22.609 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:367 @ 10/13/25 18:36:22.61 + + + + > Enter [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/13/25 18:36:22.61 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:73 @ 10/13/25 18:36:22.61 < Exit [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/13/25 18:36:22.61 (1ms) > Enter [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/13/25 18:36:22.61 < Exit [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/13/25 18:36:22.61 (0s) > Enter [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/13/25 18:36:22.61 < Exit [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/13/25 18:36:22.61 (0s) + + + > Enter [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/13/25 18:36:22.611 < Exit [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/13/25 18:36:26.14 (3.529s) > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 18:36:26.14 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 18:36:26.14 (0s) > Enter [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/13/25 18:36:26.14 STEP: Create a vmip automatically and check its binding with a lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:81 @ 10/13/25 18:36:26.14 STEP: Remove label from the lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:87 @ 10/13/25 18:36:27.26 STEP: Wait for the label to be restored by the controller - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:96 @ 10/13/25 18:36:27.393 < Exit [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/13/25 18:36:27.653 (1.513s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 18:36:27.653 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 18:36:27.653 (0s) + + + > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 18:36:27.654 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 18:36:27.654 (0s) > Enter [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/13/25 18:36:27.654 STEP: Create an intermediate vmip automatically to allocate a new ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:109 @ 10/13/25 18:36:27.654 STEP: Delete the intermediate vmip automatically and check that the lease is released - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:115 @ 10/13/25 18:36:28.037 STEP: Reuse the released lease with a static vmip - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:124 @ 10/13/25 18:36:28.295 STEP: Delete the static vmip and lease, then create another static vmip with this ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:132 @ 10/13/25 18:36:28.785 < Exit [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/13/25 18:36:29.495 (1.841s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 18:36:29.495 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 18:36:29.495 (0s) + + + + > Enter [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/13/25 18:36:29.495 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:51 @ 10/13/25 18:36:29.496 < Exit [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/13/25 18:36:29.496 (0s) > Enter [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/13/25 18:36:29.496 < Exit [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/13/25 18:36:29.496 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:82 @ 10/13/25 18:36:29.496 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:94 @ 10/13/25 18:36:29.496 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:106 @ 10/13/25 18:36:29.496 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:116 @ 10/13/25 18:36:29.496 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:132 @ 10/13/25 18:36:29.496 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:151 @ 10/13/25 18:36:29.496 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:157 @ 10/13/25 18:36:29.497 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:189 @ 10/13/25 18:36:29.497 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:216 @ 10/13/25 18:36:29.497 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:252 @ 10/13/25 18:36:29.497 + + + > Enter [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/13/25 18:36:29.497 < Exit [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/13/25 18:36:31.545 (2.048s) > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 18:36:31.545 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 18:36:31.545 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/13/25 18:36:31.545 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/13/25 18:36:39.571 (8.026s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 18:36:39.571 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 18:36:39.571 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 18:36:39.572 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 18:36:39.572 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/13/25 18:36:39.572 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:91 @ 10/13/25 18:36:39.572 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/13/25 18:36:43.901 (4.329s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 18:36:43.901 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 18:36:43.901 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1013 18:38:25.952610 45345 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: stream error: stream ID 13; INTERNAL_ERROR; received from peer\") has prevented the request from succeeding\nW1013 18:38:27.714052 45345 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 18:38:27.714188 45345 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 18:38:29.953442 45345 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 18:38:29.953472 45345 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 18:38:33.730161 45345 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 18:38:33.730250 45345 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 18:38:45.537567 45345 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 18:38:45.537627 45345 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 18:39:06.088386 45345 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 18:39:06.088520 45345 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 18:39:52.260442 45345 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 18:39:52.260476 45345 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 18:40:29.870841 45345 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 18:40:29.870910 45345 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 18:41:28.810259 45345 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 18:4... Gomega truncated this representation as it exceeds 'format.MaxLength'. Consider having the object provide a custom 'GomegaStringer' representation or adjust the parameters in Gomega's 'format' package. Learn more here: https://onsi.github.io/gomega/#adjusting-output to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:103 @ 10/13/25 18:53:26.393 + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 18:36:43.901 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 18:36:43.901 (0s) > Enter [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/13/25 18:36:43.901 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:102 @ 10/13/25 18:36:43.901 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1013 18:38:25.952610 45345 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: stream error: stream ID 13; INTERNAL_ERROR; received from peer\") has prevented the request from succeeding\nW1013 18:38:27.714052 45345 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 18:38:27.714188 45345 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 18:38:29.953442 45345 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 18:38:29.953472 45345 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 18:38:33.730161 45345 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 18:38:33.730250 45345 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 18:38:45.537567 45345 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 18:38:45.537627 45345 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 18:39:06.088386 45345 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 18:39:06.088520 45345 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 18:39:52.260442 45345 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 18:39:52.260476 45345 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 18:40:29.870841 45345 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 18:40:29.870910 45345 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 18:41:28.810259 45345 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 18:4... Gomega truncated this representation as it exceeds 'format.MaxLength'. Consider having the object provide a custom 'GomegaStringer' representation or adjust the parameters in Gomega's 'format' package. Learn more here: https://onsi.github.io/gomega/#adjusting-output to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:103 @ 10/13/25 18:53:26.393 < Exit [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/13/25 18:53:26.394 (16m42.492s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 18:53:26.394 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-sizing-policy -o yaml -l 'testcase=sizing-policy' --show-managed-fields=true error: exit status 1 stderr: E1013 18:53:27.505344 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:27.733738 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:27.967401 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:27.969054 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:27.974556 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:27.975953 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:27.980205 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:27.982958 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:27.986355 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:27.987632 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:27.991592 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:27.993310 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:27.996674 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:27.998005 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:28.230281 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:28.466740 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:28.706244 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:28.936131 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:29.172833 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:29.401763 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:29.630223 46244 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-sizing-policy -o json -l 'testcase=sizing-policy' stderr: E1013 18:53:30.400686 46246 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:30.627948 46246 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:30.869154 46246 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:30.983645 46246 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:30.988363 46246 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:30.989816 46246 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=sizing-policy", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 18:53:31.111 (4.717s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/13/25 18:53:31.112 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:126 @ 10/13/25 18:53:31.112 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:140 @ 10/13/25 18:53:31.112 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:152 @ 10/13/25 18:53:31.112 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:157 @ 10/13/25 18:53:31.112 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:163 @ 10/13/25 18:53:31.112 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:176 @ 10/13/25 18:53:31.112 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:181 @ 10/13/25 18:53:31.112 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:187 @ 10/13/25 18:53:31.113 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:202 @ 10/13/25 18:53:31.113 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:216 @ 10/13/25 18:53:31.113 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:240 @ 10/13/25 18:53:31.113 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-migration Unexpected error: <*exec.ExitError | 0x1400090c020>: exit status 1 { ProcessState: { pid: 46247, status: 256, rusage: { Utime: { Sec: 0, Usec: 38165, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 16245, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47497216, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3706, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 13, Nsignals: 55, Nvcsw: 43, Nivcsw: 678, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:42 @ 10/13/25 18:53:31.832 + > Enter [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/13/25 18:53:31.113 [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-migration Unexpected error: <*exec.ExitError | 0x1400090c020>: exit status 1 { ProcessState: { pid: 46247, status: 256, rusage: { Utime: { Sec: 0, Usec: 38165, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 16245, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47497216, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3706, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 13, Nsignals: 55, Nvcsw: 43, Nivcsw: 678, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:42 @ 10/13/25 18:53:31.832 < Exit [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/13/25 18:53:31.832 (719ms) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 18:53:31.832 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-migration -o yaml -l 'testcase=vm-migration' --show-managed-fields=true error: exit status 1 stderr: E1013 18:53:32.739847 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:32.997213 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.265722 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.501109 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.735702 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.964565 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.969807 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.971258 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.976074 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.977625 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.980436 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.981763 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.984417 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.985781 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.988451 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.989818 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.992890 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.994205 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.995454 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.996801 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:33.998130 46248 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-migration -o json -l 'testcase=vm-migration' stderr: E1013 18:53:34.629847 46249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:34.861201 46249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:34.989811 46249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:34.991436 46249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:34.996767 46249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:34.998216 46249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=vm-migration", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 18:53:35.15 (3.317s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/13/25 18:53:35.15 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:86 @ 10/13/25 18:53:35.15 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:100 @ 10/13/25 18:53:35.15 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:116 @ 10/13/25 18:53:35.15 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:131 @ 10/13/25 18:53:35.15 + + + + > Enter [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/13/25 18:53:35.151 [SKIPPED] Module SDN is disabled. Skipping all tests for module SDN. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:50 @ 10/13/25 18:53:35.891 < Exit [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/13/25 18:53:35.891 (741ms) > Enter [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/13/25 18:53:35.891 < Exit [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/13/25 18:53:35.891 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/13/25 18:53:35.891 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/13/25 18:53:35.891 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:111 @ 10/13/25 18:53:35.892 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:125 @ 10/13/25 18:53:35.892 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:141 @ 10/13/25 18:53:35.892 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:156 @ 10/13/25 18:53:35.892 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:169 @ 10/13/25 18:53:35.892 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-evacuation Unexpected error: <*exec.ExitError | 0x140006c9240>: exit status 1 { ProcessState: { pid: 46251, status: 256, rusage: { Utime: { Sec: 0, Usec: 36507, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 16774, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46891008, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3673, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 48, Nvcsw: 42, Nivcsw: 509, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:48 @ 10/13/25 18:53:36.595 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/13/25 18:53:35.892 [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-evacuation Unexpected error: <*exec.ExitError | 0x140006c9240>: exit status 1 { ProcessState: { pid: 46251, status: 256, rusage: { Utime: { Sec: 0, Usec: 36507, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 16774, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46891008, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3673, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 48, Nvcsw: 42, Nivcsw: 509, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:48 @ 10/13/25 18:53:36.595 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/13/25 18:53:36.595 (703ms) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/13/25 18:53:36.595 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-evacuation -o yaml -l 'testcase=vm-evacuation' --show-managed-fields=true error: exit status 1 stderr: E1013 18:53:37.530925 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:37.758904 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:37.992872 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:37.994602 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:38.227732 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:38.455978 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:38.690330 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:38.930057 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:39.163578 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:39.393445 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:39.625820 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:39.876495 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:39.996631 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:39.998220 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:40.233597 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:40.462803 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:40.695016 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:40.924509 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:41.152218 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:41.381108 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:41.625652 46252 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-evacuation -o json -l 'testcase=vm-evacuation' stderr: E1013 18:53:42.516840 46255 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:42.741866 46255 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:42.975095 46255 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:42.977902 46255 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:42.982966 46255 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:42.984342 46255 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=vm-evacuation", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 18:53:43.105 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 18:53:49.544 (6.439s) [FAILED] cannot delete test case resources cmd: kubectl delete virtualmachineoperations.virtualization.deckhouse.io --namespace head-5073ae15-end-to-end-vm-evacuation -l 'testcase=vm-evacuation' stderr: E1013 18:53:47.171751 46258 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:47.409002 46258 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:47.646436 46258 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:47.884311 46258 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:48.119913 46258 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:48.359764 46258 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:48.594335 46258 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:48.833816 46258 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:49.076723 46258 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:49.306952 46258 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:49.537380 46258 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Unexpected error: <*exec.ExitError | 0x14000711320>: exit status 1 { ProcessState: { pid: 46258, status: 256, rusage: { Utime: { Sec: 0, Usec: 99235, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 95745, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 51494912, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3959, Majflt: 4, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 33, Msgrcv: 35, Nsignals: 81, Nvcsw: 40, Nivcsw: 2397, }, }, Stderr: nil, } occurred In [AfterEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:626 @ 10/13/25 18:53:49.544 < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/13/25 18:53:49.544 (12.949s) + + + + > Enter [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/13/25 18:53:49.544 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:41 @ 10/13/25 18:53:49.544 < Exit [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/13/25 18:53:49.545 (0s) > Enter [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/13/25 18:53:49.545 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 18:53:49.545 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 18:53:56.386 (6.842s) [FAILED] cannot delete test case resources cmd: kubectl delete virtualmachineoperations.virtualization.deckhouse.io -l 'testcase=vm-migration-cancel' stderr: E1013 18:53:54.058335 46288 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:54.288585 46288 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:54.524370 46288 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:54.760657 46288 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:54.999136 46288 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:55.226544 46288 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:55.452782 46288 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:55.681540 46288 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:55.917561 46288 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:56.147628 46288 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:56.378801 46288 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Unexpected error: <*exec.ExitError | 0x140002b6000>: exit status 1 { ProcessState: { pid: 46288, status: 256, rusage: { Utime: { Sec: 0, Usec: 94359, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 91409, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 51642368, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3962, Majflt: 8, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 33, Msgrcv: 35, Nsignals: 113, Nvcsw: 98, Nivcsw: 2653, }, }, Stderr: nil, } occurred In [AfterEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:626 @ 10/13/25 18:53:56.387 < Exit [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/13/25 18:53:56.387 (6.843s) + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-images-creation Unexpected error: <*exec.ExitError | 0x140007102c0>: exit status 1 { ProcessState: { pid: 46302, status: 256, rusage: { Utime: { Sec: 0, Usec: 33620, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 14721, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47398912, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3704, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 11, Nsignals: 38, Nvcsw: 26, Nivcsw: 527, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:49 @ 10/13/25 18:53:57.282 + > Enter [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/13/25 18:53:56.387 [FAILED] kubectl create namespace head-5073ae15-end-to-end-images-creation Unexpected error: <*exec.ExitError | 0x140007102c0>: exit status 1 { ProcessState: { pid: 46302, status: 256, rusage: { Utime: { Sec: 0, Usec: 33620, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 14721, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47398912, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3704, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 11, Nsignals: 38, Nvcsw: 26, Nivcsw: 527, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:49 @ 10/13/25 18:53:57.282 < Exit [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/13/25 18:53:57.282 (895ms) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/13/25 18:53:57.282 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-images-creation -o yaml -l 'testcase=images-creation' --show-managed-fields=true error: exit status 1 stderr: E1013 18:53:58.012567 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:58.263613 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:58.495812 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:58.722532 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:58.949368 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:59.200603 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:59.428876 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:59.657924 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:53:59.894630 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:00.126823 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:00.369902 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:00.610518 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:00.841484 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:00.953968 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:00.958630 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:00.960059 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:00.964739 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:00.966152 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:00.967521 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:00.968834 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:00.970192 46311 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-images-creation -o json -l 'testcase=images-creation' stderr: E1013 18:54:01.832659 46313 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:02.063006 46313 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:02.296688 46313 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:02.523312 46313 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:02.762425 46313 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:02.992313 46313 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=images-creation", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/13/25 18:54:03.114 (5.832s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:94 @ 10/13/25 18:54:03.115 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:103 @ 10/13/25 18:54:03.115 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:114 @ 10/13/25 18:54:03.115 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:128 @ 10/13/25 18:54:03.115 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:139 @ 10/13/25 18:54:03.115 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-restore-safe Unexpected error: <*exec.ExitError | 0x14000711520>: exit status 1 { ProcessState: { pid: 46318, status: 256, rusage: { Utime: { Sec: 0, Usec: 35643, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 19570, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46743552, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3662, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 47, Nvcsw: 43, Nivcsw: 541, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:58 @ 10/13/25 18:54:03.634 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/13/25 18:54:03.115 [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-restore-safe Unexpected error: <*exec.ExitError | 0x14000711520>: exit status 1 { ProcessState: { pid: 46318, status: 256, rusage: { Utime: { Sec: 0, Usec: 35643, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 19570, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46743552, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3662, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 47, Nvcsw: 43, Nivcsw: 541, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:58 @ 10/13/25 18:54:03.634 < Exit [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/13/25 18:54:03.634 (519ms) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 18:54:03.634 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-restore-safe -o yaml -l 'testcase=vm-restore-safe' --show-managed-fields=true error: exit status 1 stderr: E1013 18:54:04.711780 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:04.942554 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:05.176141 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:05.412064 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:05.661583 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:05.893286 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:06.135970 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:06.364473 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:06.598678 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:06.848923 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:06.969662 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:06.971195 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:06.976983 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:06.978301 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:06.981689 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:06.983249 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:06.987871 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:06.989207 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:06.990664 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:06.992092 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:06.993426 46319 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-restore-safe -o json -l 'testcase=vm-restore-safe' stderr: E1013 18:54:07.871195 46321 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:07.990986 46321 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:07.995532 46321 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:07.996812 46321 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:08.232369 46321 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:08.462410 46321 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=vm-restore-safe", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. [PANICKED] Test Panicked In [AfterEach] at: /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/panic.go:262 @ 10/13/25 18:54:08.592 runtime error: invalid memory address or nil pointer dereference Full Stack Trace github.com/deckhouse/virtualization/tests/e2e.init.func23.3() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:70 +0xe4 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 18:54:08.592 (4.958s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/13/25 18:54:08.593 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:115 @ 10/13/25 18:54:08.593 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:158 @ 10/13/25 18:54:08.593 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:348 @ 10/13/25 18:54:08.593 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:374 @ 10/13/25 18:54:08.593 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-affinity-toleration Unexpected error: <*exec.ExitError | 0x140002b66e0>: exit status 1 { ProcessState: { pid: 46322, status: 256, rusage: { Utime: { Sec: 0, Usec: 35187, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 19285, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46612480, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3660, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 13, Nsignals: 42, Nvcsw: 48, Nivcsw: 582, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:69 @ 10/13/25 18:54:09.168 + > Enter [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/13/25 18:54:08.593 [FAILED] kubectl create namespace head-5073ae15-end-to-end-affinity-toleration Unexpected error: <*exec.ExitError | 0x140002b66e0>: exit status 1 { ProcessState: { pid: 46322, status: 256, rusage: { Utime: { Sec: 0, Usec: 35187, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 19285, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46612480, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3660, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 13, Nsignals: 42, Nvcsw: 48, Nivcsw: 582, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:69 @ 10/13/25 18:54:09.168 < Exit [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/13/25 18:54:09.169 (576ms) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 18:54:09.169 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-affinity-toleration -o yaml -l 'testcase=affinity-toleration' --show-managed-fields=true error: exit status 1 stderr: E1013 18:54:10.227936 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:10.454047 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:10.692246 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:10.917684 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:11.147349 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:11.373017 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:11.604651 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:11.830717 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:12.075152 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:12.345778 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:12.579803 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:12.811871 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:13.043382 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:13.269797 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:13.511794 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:13.739386 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:13.968536 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:13.970306 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:13.971944 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:13.973502 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:13.975042 46323 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-affinity-toleration -o json -l 'testcase=affinity-toleration' stderr: E1013 18:54:14.961279 46324 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:15.193750 46324 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:15.428084 46324 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:15.663381 46324 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:15.897079 46324 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:16.129949 46324 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=affinity-toleration", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 18:54:16.255 (7.087s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/13/25 18:54:16.256 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:120 @ 10/13/25 18:54:16.256 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:283 @ 10/13/25 18:54:16.256 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:372 @ 10/13/25 18:54:16.256 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:467 @ 10/13/25 18:54:16.256 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-complex-test Unexpected error: <*exec.ExitError | 0x1400090ce40>: exit status 1 { ProcessState: { pid: 46326, status: 256, rusage: { Utime: { Sec: 0, Usec: 38453, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 21237, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47349760, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3702, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 61, Nvcsw: 42, Nivcsw: 679, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:58 @ 10/13/25 18:54:17.089 + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/13/25 18:54:16.256 [FAILED] kubectl create namespace head-5073ae15-end-to-end-complex-test Unexpected error: <*exec.ExitError | 0x1400090ce40>: exit status 1 { ProcessState: { pid: 46326, status: 256, rusage: { Utime: { Sec: 0, Usec: 38453, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 21237, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47349760, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3702, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 61, Nvcsw: 42, Nivcsw: 679, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:58 @ 10/13/25 18:54:17.089 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/13/25 18:54:17.089 (832ms) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 18:54:17.089 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-complex-test -o yaml -l 'testcase=complex-test' --show-managed-fields=true error: exit status 1 stderr: E1013 18:54:17.858739 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:17.981790 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:17.988376 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:17.989661 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:17.993376 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:17.995114 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:17.998767 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:18.225815 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:18.462311 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:18.696126 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:18.931111 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:19.163758 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:19.398599 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:19.634165 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:19.867519 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:19.985040 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:19.991349 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:19.992814 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:19.994238 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:19.995660 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:19.997044 46327 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-complex-test -o json -l 'testcase=complex-test' stderr: E1013 18:54:20.968662 46330 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:20.970540 46330 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:20.977367 46330 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:20.979913 46330 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:20.982904 46330 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:20.984152 46330 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=complex-test", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 18:54:21.102 (4.014s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/13/25 18:54:21.103 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/13/25 18:54:21.103 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/13/25 18:54:21.103 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/13/25 18:54:21.103 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/13/25 18:54:21.103 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:141 @ 10/13/25 18:54:21.103 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:152 @ 10/13/25 18:54:21.103 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:162 @ 10/13/25 18:54:21.104 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:173 @ 10/13/25 18:54:21.104 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:184 @ 10/13/25 18:54:21.104 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:196 @ 10/13/25 18:54:21.104 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:220 @ 10/13/25 18:54:21.104 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:245 @ 10/13/25 18:54:21.104 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:268 @ 10/13/25 18:54:21.104 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:280 @ 10/13/25 18:54:21.104 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:299 @ 10/13/25 18:54:21.104 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:316 @ 10/13/25 18:54:21.104 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:341 @ 10/13/25 18:54:21.105 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:354 @ 10/13/25 18:54:21.105 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:358 @ 10/13/25 18:54:21.105 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:375 @ 10/13/25 18:54:21.105 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:388 @ 10/13/25 18:54:21.105 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:405 @ 10/13/25 18:54:21.105 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:444 @ 10/13/25 18:54:21.105 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:479 @ 10/13/25 18:54:21.105 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:497 @ 10/13/25 18:54:21.105 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:512 @ 10/13/25 18:54:21.105 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:528 @ 10/13/25 18:54:21.105 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:545 @ 10/13/25 18:54:21.105 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-connectivity Unexpected error: <*exec.ExitError | 0x14000b14fc0>: exit status 1 { ProcessState: { pid: 46331, status: 256, rusage: { Utime: { Sec: 0, Usec: 35689, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 19281, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47824896, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3730, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 30, Nvcsw: 15, Nivcsw: 583, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:63 @ 10/13/25 18:54:21.768 + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/13/25 18:54:21.106 [FAILED] kubectl create namespace head-5073ae15-end-to-end-connectivity Unexpected error: <*exec.ExitError | 0x14000b14fc0>: exit status 1 { ProcessState: { pid: 46331, status: 256, rusage: { Utime: { Sec: 0, Usec: 35689, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 19281, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47824896, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3730, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 30, Nvcsw: 15, Nivcsw: 583, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:63 @ 10/13/25 18:54:21.768 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/13/25 18:54:21.768 (663ms) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 18:54:21.768 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-connectivity -o yaml -l 'testcase=vm-connectivity' --show-managed-fields=true error: exit status 1 stderr: E1013 18:54:22.727459 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:22.956880 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:23.191194 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:23.424356 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:23.692241 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:23.954342 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:24.206321 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:24.438500 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:24.712313 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:24.946464 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:25.176670 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:25.425202 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:25.697674 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:25.921017 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:26.164039 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:26.388079 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:26.616389 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:26.848282 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:26.961830 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:26.963392 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:26.964837 46334 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-connectivity -o json -l 'testcase=vm-connectivity' stderr: E1013 18:54:27.598994 46336 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:27.837104 46336 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:28.076194 46336 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:28.305230 46336 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:28.539165 46336 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:28.769068 46336 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=vm-connectivity", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 18:54:28.895 (7.127s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/13/25 18:54:28.895 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/13/25 18:54:28.895 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:118 @ 10/13/25 18:54:28.896 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:129 @ 10/13/25 18:54:28.896 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:142 @ 10/13/25 18:54:28.896 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:166 @ 10/13/25 18:54:28.896 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:174 @ 10/13/25 18:54:28.896 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:179 @ 10/13/25 18:54:28.896 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:187 @ 10/13/25 18:54:28.896 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:198 @ 10/13/25 18:54:28.896 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:209 @ 10/13/25 18:54:28.896 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:222 @ 10/13/25 18:54:28.896 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:228 @ 10/13/25 18:54:28.896 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:240 @ 10/13/25 18:54:28.896 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:250 @ 10/13/25 18:54:28.896 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:258 @ 10/13/25 18:54:28.896 + + + [FAILED] error: error validating "/tmp/testdata/importer-network-policy/project": error validating data: failed to download openapi: the server could not find the requested resource; if you choose to ignore these errors, turn validation off with --validate=false Expected <bool>: false to equal <bool>: true In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:69 @ 10/13/25 18:54:29.67 + > Enter [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/13/25 18:54:28.897 < Exit [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/13/25 18:54:28.898 (1ms) > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 18:54:28.898 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 18:54:28.898 (0s) > Enter [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/13/25 18:54:28.898 [FAILED] error: error validating "/tmp/testdata/importer-network-policy/project": error validating data: failed to download openapi: the server could not find the requested resource; if you choose to ignore these errors, turn validation off with --validate=false Expected <bool>: false to equal <bool>: true In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:69 @ 10/13/25 18:54:29.67 < Exit [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/13/25 18:54:29.67 (771ms) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 18:54:29.67 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-importer-network-policy -o yaml -l 'testcase=importer-network-policy' --show-managed-fields=true error: exit status 1 stderr: E1013 18:54:30.379771 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:30.612803 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:30.852497 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:30.968281 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:30.972162 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:30.973372 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:30.978443 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:30.979895 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:30.983142 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:30.984726 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:30.988061 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:30.989367 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:30.993513 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:30.994886 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:30.998088 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:30.999505 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:31.231509 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:31.461442 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:31.691644 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:31.921010 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:32.153547 46339 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-importer-network-policy -o json -l 'testcase=importer-network-policy' stderr: E1013 18:54:32.900332 46340 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:33.126004 46340 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:33.355833 46340 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:33.580827 46340 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:33.811664 46340 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:34.040895 46340 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=importer-network-policy", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 18:54:34.162 (4.492s) > Enter [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/13/25 18:54:34.162 STEP: Delete manifests - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:50 @ 10/13/25 18:54:34.162 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 18:54:34.162 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 18:54:36.296 (2.134s) < Exit [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/13/25 18:54:36.296 (2.134s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/13/25 18:54:36.297 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/13/25 18:54:36.297 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/13/25 18:54:36.297 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/13/25 18:54:36.297 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/13/25 18:54:36.297 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-disk-attachment Unexpected error: <*exec.ExitError | 0x140002b6520>: exit status 1 { ProcessState: { pid: 46343, status: 256, rusage: { Utime: { Sec: 0, Usec: 38283, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 17772, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47382528, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3705, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 13, Nsignals: 51, Nvcsw: 80, Nivcsw: 654, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:61 @ 10/13/25 18:54:37.205 + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/13/25 18:54:36.297 [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-disk-attachment Unexpected error: <*exec.ExitError | 0x140002b6520>: exit status 1 { ProcessState: { pid: 46343, status: 256, rusage: { Utime: { Sec: 0, Usec: 38283, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 17772, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47382528, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3705, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 13, Nsignals: 51, Nvcsw: 80, Nivcsw: 654, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:61 @ 10/13/25 18:54:37.205 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/13/25 18:54:37.205 (908ms) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 18:54:37.205 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-disk-attachment -o yaml -l 'testcase=vm-disk-attachment' --show-managed-fields=true error: exit status 1 stderr: E1013 18:54:37.911685 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.143148 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.377485 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.608312 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.848034 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.961656 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.966076 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.967615 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.971464 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.972778 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.974819 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.976479 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.979087 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.980402 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.982940 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.984242 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.987284 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.988632 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.989948 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.991269 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:38.992575 46344 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-disk-attachment -o json -l 'testcase=vm-disk-attachment' stderr: E1013 18:54:39.625368 46346 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:39.849556 46346 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:39.965858 46346 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:39.967564 46346 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:39.973651 46346 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:39.975563 46346 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=vm-disk-attachment", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 18:54:40.099 (2.894s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/13/25 18:54:40.099 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/13/25 18:54:40.099 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/13/25 18:54:40.099 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/13/25 18:54:40.099 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/13/25 18:54:40.099 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/13/25 18:54:40.099 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/13/25 18:54:40.099 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/13/25 18:54:40.1 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/13/25 18:54:40.1 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/13/25 18:54:40.1 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/13/25 18:54:40.1 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/13/25 18:54:40.1 + + + [FAILED] Unexpected error: <*exec.ExitError | 0x140000ef2c0>: exit status 1 { ProcessState: { pid: 46348, status: 256, rusage: { Utime: { Sec: 0, Usec: 90515, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 60392, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 49807360, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3856, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 21, Msgrcv: 21, Nsignals: 114, Nvcsw: 90, Nivcsw: 1842, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:65 @ 10/13/25 18:54:41.995 + > Enter [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/13/25 18:54:40.1 [FAILED] Unexpected error: <*exec.ExitError | 0x140000ef2c0>: exit status 1 { ProcessState: { pid: 46348, status: 256, rusage: { Utime: { Sec: 0, Usec: 90515, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 60392, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 49807360, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3856, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 21, Msgrcv: 21, Nsignals: 114, Nvcsw: 90, Nivcsw: 1842, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:65 @ 10/13/25 18:54:41.995 < Exit [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/13/25 18:54:41.995 (1.895s) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/13/25 18:54:41.995 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-image-hotplug -o yaml -l 'testcase=image-hotplug' --show-managed-fields=true error: exit status 1 stderr: E1013 18:54:42.653914 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:42.918331 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.164198 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.390804 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.621591 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.847034 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.965989 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.967627 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.972564 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.973968 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.979667 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.981167 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.984717 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.986215 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.991229 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.992492 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.995656 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.996982 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.998274 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:43.999662 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:44.224268 46349 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-image-hotplug -o json -l 'testcase=image-hotplug' stderr: E1013 18:54:44.940352 46350 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:45.174324 46350 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:45.414446 46350 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:45.645865 46350 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:45.880937 46350 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:46.114692 46350 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=image-hotplug", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/13/25 18:54:46.238 (4.244s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:112 @ 10/13/25 18:54:46.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:147 @ 10/13/25 18:54:46.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:190 @ 10/13/25 18:54:46.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:196 @ 10/13/25 18:54:46.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:204 @ 10/13/25 18:54:46.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:224 @ 10/13/25 18:54:46.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:236 @ 10/13/25 18:54:46.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:258 @ 10/13/25 18:54:46.24 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:284 @ 10/13/25 18:54:46.24 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:294 @ 10/13/25 18:54:46.24 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:309 @ 10/13/25 18:54:46.24 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-label-annotation Unexpected error: <*exec.ExitError | 0x1400090c380>: exit status 1 { ProcessState: { pid: 46352, status: 256, rusage: { Utime: { Sec: 0, Usec: 36050, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 19629, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46989312, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3675, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 14, Nsignals: 36, Nvcsw: 25, Nivcsw: 620, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:48 @ 10/13/25 18:54:46.835 + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/13/25 18:54:46.24 [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-label-annotation Unexpected error: <*exec.ExitError | 0x1400090c380>: exit status 1 { ProcessState: { pid: 46352, status: 256, rusage: { Utime: { Sec: 0, Usec: 36050, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 19629, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46989312, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3675, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 14, Nsignals: 36, Nvcsw: 25, Nivcsw: 620, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:48 @ 10/13/25 18:54:46.835 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/13/25 18:54:46.836 (596ms) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 18:54:46.836 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-label-annotation -o yaml -l 'testcase=vm-label-annotation' --show-managed-fields=true error: exit status 1 stderr: E1013 18:54:47.869391 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:47.983959 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:47.989837 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:47.991358 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:47.996512 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:47.997727 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:48.227860 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:48.453240 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:48.707504 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:48.955041 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:49.188249 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:49.415590 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:49.652257 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:49.882352 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:50.114852 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:50.342986 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:50.575275 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:50.800664 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:51.038130 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:51.264347 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:51.489648 46353 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-label-annotation -o json -l 'testcase=vm-label-annotation' stderr: E1013 18:54:52.233910 46354 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:52.485413 46354 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:52.724512 46354 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:52.959130 46354 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:53.193024 46354 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 18:54:53.423917 46354 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=vm-label-annotation", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 18:54:53.552 (6.716s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/13/25 18:54:53.552 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/13/25 18:54:53.552 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/13/25 18:54:53.552 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/13/25 18:54:53.552 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/13/25 18:54:53.552 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/13/25 18:54:53.553 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/13/25 18:54:53.553 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/13/25 18:54:53.553 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/13/25 18:54:53.553 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/13/25 18:54:53.553 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/13/25 18:54:53.553 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/13/25 18:54:53.553 + + + [FAILED] Expected success, but got an error: <*errors.StatusError | 0x14000627d60>: the server could not find the requested resource (get pods) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "the server could not find the requested resource (get pods)", Reason: "NotFound", Details: { Name: "", Group: "", Kind: "pods", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>404 Not Found</title></head>\r\n<body>\r\n<center><h1>404 Not Found</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 404, }, } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/13/25 18:54:54.122 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 18:54:53.553 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 18:54:53.553 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 18:54:53.553 [FAILED] Expected success, but got an error: <*errors.StatusError | 0x14000627d60>: the server could not find the requested resource (get pods) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "the server could not find the requested resource (get pods)", Reason: "NotFound", Details: { Name: "", Group: "", Kind: "pods", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>404 Not Found</title></head>\r\n<body>\r\n<center><h1>404 Not Found</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 404, }, } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/13/25 18:54:54.122 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 18:54:54.122 (569ms) + + + [FAILED] Expected success, but got an error: <*errors.StatusError | 0x14000033900>: the server could not find the requested resource (get projects.deckhouse.io) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "the server could not find the requested resource (get projects.deckhouse.io)", Reason: "NotFound", Details: { Name: "", Group: "deckhouse.io", Kind: "projects", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>404 Not Found</title></head>\r\n<body>\r\n<center><h1>404 Not Found</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 404, }, } In [DeferCleanup (Suite)] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:177 @ 10/13/25 18:54:54.235 + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/13/25 18:54:54.123 [FAILED] Expected success, but got an error: <*errors.StatusError | 0x14000033900>: the server could not find the requested resource (get projects.deckhouse.io) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "the server could not find the requested resource (get projects.deckhouse.io)", Reason: "NotFound", Details: { Name: "", Group: "deckhouse.io", Kind: "projects", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>404 Not Found</title></head>\r\n<body>\r\n<center><h1>404 Not Found</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 404, }, } In [DeferCleanup (Suite)] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:177 @ 10/13/25 18:54:54.235 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/13/25 18:54:54.235 (112ms) + + + \ No newline at end of file diff --git a/artifacts/clean-cephrbd-20251013-183813-18792/junit.xml b/artifacts/clean-cephrbd-20251013-183813-18792/junit.xml new file mode 100644 index 0000000000..2449804c4f --- /dev/null +++ b/artifacts/clean-cephrbd-20251013-183813-18792/junit.xml @@ -0,0 +1,852 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 18:59:43.973 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 18:59:48.805 (4.832s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 18:59:48.805 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 18:59:48.805 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.807 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:59:48.807 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.807 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.807 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:59:48.807 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.807 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.807 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:59:48.807 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.807 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.807 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:59:48.807 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.807 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.807 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:59:48.807 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.807 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.807 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:59:48.807 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.808 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.808 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:59:48.808 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.808 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.808 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:59:48.808 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.808 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.808 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:59:48.808 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.808 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.808 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:59:48.808 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.808 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.808 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:59:48.808 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.808 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.808 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 18:59:48.808 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 18:59:48.808 (0s) + + + > Enter [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/13/25 18:59:48.808 < Exit [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/13/25 18:59:50.738 (1.929s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/13/25 18:59:50.738 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/13/25 19:00:00.257 (9.52s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 19:00:00.257 < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 19:00:00.257 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:1, cap:6>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-node-selector --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-node-selector\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:110 @ 10/13/25 19:17:57.219 + > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/13/25 19:00:00.258 STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/13/25 19:00:00.258 END STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/13/25 19:00:10.917 (10.66s) STEP: `VirtualMachineClasses` should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:95 @ 10/13/25 19:00:10.917 END STEP: `VirtualMachineClasses` should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:95 @ 10/13/25 19:00:13.666 (2.749s) STEP: `VirtualDisks` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:102 @ 10/13/25 19:00:13.666 END STEP: `VirtualDisks` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:102 @ 10/13/25 19:01:15.09 (1m1.424s) STEP: `VirtualMachines` agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:109 @ 10/13/25 19:01:15.09 END STEP: `VirtualMachines` agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:109 @ 10/13/25 19:17:57.218 (16m42.106s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:1, cap:6>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-node-selector --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-node-selector\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:110 @ 10/13/25 19:17:57.219 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/13/25 19:17:57.22 (17m56.94s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 19:17:57.22 Failed to get logs: PodName: virt-launcher-head-5073ae15-vm-node-selector-vm4dk Error: Error from server (BadRequest): container "d8v-compute" in pod "virt-launcher-head-5073ae15-vm-node-selector-vm4dk" is waiting to start: ContainerCreating < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 19:18:18.811 (21.591s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:120 @ 10/13/25 19:18:18.811 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:283 @ 10/13/25 19:18:18.812 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:372 @ 10/13/25 19:18:18.812 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:467 @ 10/13/25 19:18:18.812 + + + [FAILED] Warning: resource namespaces/head-5073ae15-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "linstor-thin-r2" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x140005a02a0>: exit status 1 { ProcessState: { pid: 47474, status: 256, rusage: { Utime: { Sec: 0, Usec: 299823, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 64832, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 69861376, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5069, Majflt: 7, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 49, Msgrcv: 54, Nsignals: 333, Nvcsw: 332, Nivcsw: 2814, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/13/25 19:18:26.222 + > Enter [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/13/25 19:18:18.812 < Exit [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/13/25 19:18:21.852 (3.04s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/13/25 19:18:21.852 [FAILED] Warning: resource namespaces/head-5073ae15-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "linstor-thin-r2" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x140005a02a0>: exit status 1 { ProcessState: { pid: 47474, status: 256, rusage: { Utime: { Sec: 0, Usec: 299823, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 64832, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 69861376, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5069, Majflt: 7, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 49, Msgrcv: 54, Nsignals: 333, Nvcsw: 332, Nivcsw: 2814, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/13/25 19:18:26.222 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/13/25 19:18:26.222 (4.37s) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/13/25 19:18:26.222 The list of pods is empty; nothing to dump. < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/13/25 19:18:32.148 (5.926s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:112 @ 10/13/25 19:18:32.148 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:147 @ 10/13/25 19:18:32.149 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:190 @ 10/13/25 19:18:32.149 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:196 @ 10/13/25 19:18:32.149 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:204 @ 10/13/25 19:18:32.149 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:224 @ 10/13/25 19:18:32.149 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:236 @ 10/13/25 19:18:32.149 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:258 @ 10/13/25 19:18:32.149 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:284 @ 10/13/25 19:18:32.149 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:294 @ 10/13/25 19:18:32.149 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:309 @ 10/13/25 19:18:32.149 + + + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/13/25 19:18:32.149 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/13/25 19:18:34.117 (1.968s) > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:18:34.117 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:18:34.117 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/13/25 19:18:34.117 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/13/25 19:18:36.96 (2.842s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:18:36.96 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:18:36.96 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:18:36.96 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:18:36.96 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/13/25 19:18:36.96 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:75 @ 10/13/25 19:18:36.96 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/13/25 19:18:42.654 (5.694s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:18:42.654 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:18:42.654 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:18:42.655 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:18:42.655 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/13/25 19:18:42.655 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 @ 10/13/25 19:18:42.655 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/13/25 19:21:20.29 (2m37.616s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:20.291 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:20.291 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:20.291 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:20.291 (0s) > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/13/25 19:21:20.291 STEP: Virtual machine phase should be Running - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:97 @ 10/13/25 19:21:20.291 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/13/25 19:21:23.432 (3.141s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:23.432 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:23.432 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:23.432 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:23.432 (0s) > Enter [It] marks VMs with label map["specialKey":"specialValue"] - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/13/25 19:21:23.432 < Exit [It] marks VMs with label map["specialKey":"specialValue"] - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/13/25 19:21:26.079 (2.646s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:26.079 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:26.079 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:26.079 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:26.079 (0s) > Enter [It] checks VMs and pods labels after VMs labeling - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/13/25 19:21:26.079 < Exit [It] checks VMs and pods labels after VMs labeling - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/13/25 19:21:28.387 (2.308s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:28.387 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:28.387 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:28.387 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:28.387 (0s) > Enter [It] removes label map[specialKey:specialValue] from VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/13/25 19:21:28.387 < Exit [It] removes label map[specialKey:specialValue] from VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/13/25 19:21:31.065 (2.677s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:31.065 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:31.065 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:31.065 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:31.065 (0s) > Enter [It] checks VMs and pods labels after VMs unlabeling - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/13/25 19:21:31.065 < Exit [It] checks VMs and pods labels after VMs unlabeling - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/13/25 19:21:33.324 (2.259s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:33.324 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:33.324 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:33.324 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:33.324 (0s) > Enter [It] marks VMs with annotation map["specialKey":"specialValue"] - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/13/25 19:21:33.324 < Exit [It] marks VMs with annotation map["specialKey":"specialValue"] - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/13/25 19:21:36.006 (2.682s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:36.006 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:36.006 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:36.007 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:36.007 (0s) > Enter [It] checks VMs and pods annotations after VMs annotating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/13/25 19:21:36.007 < Exit [It] checks VMs and pods annotations after VMs annotating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/13/25 19:21:38.26 (2.254s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:38.26 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:38.26 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:38.261 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:38.261 (0s) > Enter [It] removes annotation map[specialKey:specialValue] from VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/13/25 19:21:38.261 < Exit [It] removes annotation map[specialKey:specialValue] from VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/13/25 19:21:40.902 (2.641s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:40.902 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:40.902 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:40.902 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:40.902 (0s) > Enter [It] checks VMs and pods annotations after VMs unannotating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/13/25 19:21:40.902 < Exit [It] checks VMs and pods annotations after VMs unannotating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/13/25 19:21:42.891 (1.988s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:42.891 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:21:42.891 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:42.891 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:21:42.891 (0s) > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/13/25 19:21:42.891 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 19:21:42.891 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 19:22:39.698 (56.807s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/13/25 19:22:39.698 (56.807s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:22:39.698 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:22:39.698 (0s) + + + > Enter [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/13/25 19:22:39.699 < Exit [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/13/25 19:22:41.686 (1.987s) > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 19:22:41.686 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 19:22:41.686 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/13/25 19:22:41.686 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/13/25 19:22:44.219 (2.533s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 19:22:44.219 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 19:22:44.22 (1ms) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 19:22:44.221 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 19:22:44.221 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/13/25 19:22:44.221 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:68 @ 10/13/25 19:22:44.221 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/13/25 19:22:56.586 (12.365s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 19:22:56.586 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 19:22:56.586 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 19:22:56.587 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 19:22:56.587 (0s) > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/13/25 19:22:56.587 STEP: VM should be in Running phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:79 @ 10/13/25 19:22:56.587 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/13/25 19:22:59.696 (3.109s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 19:22:59.696 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 19:22:59.696 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 19:22:59.696 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 19:22:59.696 (0s) > Enter [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/13/25 19:22:59.696 < Exit [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/13/25 19:22:59.696 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 19:22:59.696 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 19:22:59.696 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 19:22:59.696 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 19:22:59.696 (0s) > Enter [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/13/25 19:22:59.696 < Exit [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/13/25 19:22:59.696 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 19:22:59.696 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 19:22:59.696 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 19:22:59.696 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 19:22:59.696 (0s) > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/13/25 19:22:59.696 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 19:22:59.696 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 19:24:09.755 (1m10.059s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/13/25 19:24:09.755 (1m10.059s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 19:24:09.755 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 19:24:09.755 (0s) + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:51 @ 10/13/25 19:24:11.657 + > Enter [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/13/25 19:24:09.755 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:51 @ 10/13/25 19:24:11.657 < Exit [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/13/25 19:24:11.658 (1.902s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/13/25 19:24:11.658 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/13/25 19:24:17.514 (5.856s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:94 @ 10/13/25 19:24:17.514 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:103 @ 10/13/25 19:24:17.514 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:114 @ 10/13/25 19:24:17.514 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:128 @ 10/13/25 19:24:17.514 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:139 @ 10/13/25 19:24:17.514 + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 @ 10/13/25 19:41:06.846 + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/13/25 19:24:17.514 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/13/25 19:24:19.006 (1.492s) > Enter [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/13/25 19:24:19.006 < Exit [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/13/25 19:24:24.654 (5.648s) > Enter [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/13/25 19:24:24.654 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 @ 10/13/25 19:24:24.654 [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 @ 10/13/25 19:41:06.846 < Exit [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/13/25 19:41:06.846 (16m42.146s) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/13/25 19:41:06.846 Failed to get logs: PodName: virt-launcher-head-5073ae15-vm-migration-bios-8wnbj Error: Error from server (BadRequest): container "d8v-compute" in pod "virt-launcher-head-5073ae15-vm-migration-bios-8wnbj" is waiting to start: ContainerCreating Failed to get logs: PodName: virt-launcher-head-5073ae15-vm-migration-uefi-m8r8g Error: Error from server (BadRequest): container "d8v-compute" in pod "virt-launcher-head-5073ae15-vm-migration-uefi-m8r8g" is waiting to start: ContainerCreating STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 19:41:18.09 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 19:41:23.934 (5.844s) < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/13/25 19:41:23.934 (17.088s) + + + > Enter [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/13/25 19:41:23.934 < Exit [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/13/25 19:41:25.934 (2s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/13/25 19:41:25.935 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/13/25 19:41:31.734 (5.8s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 19:41:31.734 < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 19:41:31.735 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:77 @ 10/13/25 19:58:13.91 + > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/13/25 19:41:31.735 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:76 @ 10/13/25 19:41:31.735 [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:77 @ 10/13/25 19:58:13.91 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/13/25 19:58:13.91 (16m42.139s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 19:58:13.91 Failed to get logs: PodName: virt-launcher-head-5073ae15-vm-migration-bios-lnjs2 Error: Error from server (BadRequest): container "d8v-compute" in pod "virt-launcher-head-5073ae15-vm-migration-bios-lnjs2" is waiting to start: ContainerCreating Failed to get logs: PodName: virt-launcher-head-5073ae15-vm-migration-uefi-vwc8b Error: Error from server (BadRequest): container "d8v-compute" in pod "virt-launcher-head-5073ae15-vm-migration-uefi-vwc8b" is waiting to start: ContainerCreating < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 19:58:24.742 (10.832s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:86 @ 10/13/25 19:58:24.743 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:100 @ 10/13/25 19:58:24.743 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:116 @ 10/13/25 19:58:24.743 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:131 @ 10/13/25 19:58:24.743 + + + > Enter [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/13/25 19:58:24.743 < Exit [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/13/25 19:58:28.016 (3.273s) > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 19:58:28.016 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 19:58:28.016 (0s) > Enter [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/13/25 19:58:28.016 STEP: Create a vmip automatically and check its binding with a lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:81 @ 10/13/25 19:58:28.016 STEP: Remove label from the lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:87 @ 10/13/25 19:58:28.907 STEP: Wait for the label to be restored by the controller - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:96 @ 10/13/25 19:58:29.03 < Exit [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/13/25 19:58:29.265 (1.249s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 19:58:29.265 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 19:58:29.265 (0s) + + + > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 19:58:29.265 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 19:58:29.265 (0s) > Enter [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/13/25 19:58:29.265 STEP: Create an intermediate vmip automatically to allocate a new ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:109 @ 10/13/25 19:58:29.265 STEP: Delete the intermediate vmip automatically and check that the lease is released - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:115 @ 10/13/25 19:58:29.623 STEP: Reuse the released lease with a static vmip - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:124 @ 10/13/25 19:58:29.859 STEP: Delete the static vmip and lease, then create another static vmip with this ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:132 @ 10/13/25 19:58:30.227 < Exit [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/13/25 19:58:30.888 (1.623s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 19:58:30.888 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 19:58:30.888 (0s) + + + > Enter [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/13/25 19:58:30.889 < Exit [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/13/25 19:58:30.889 (1ms) > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:58:30.889 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:58:30.889 (0s) > Enter [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/13/25 19:58:30.889 < Exit [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/13/25 19:58:32.438 (1.549s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:58:32.438 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:58:32.438 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:58:32.438 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:58:32.438 (0s) > Enter [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/13/25 19:58:32.438 STEP: Project should be deployed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:73 @ 10/13/25 19:58:32.438 < Exit [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/13/25 19:58:35.172 (2.733s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:58:35.172 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:58:35.172 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:58:35.172 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:58:35.172 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/13/25 19:58:35.172 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/13/25 19:58:38.496 (3.324s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:58:38.496 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:58:38.496 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:58:38.496 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:58:38.497 (0s) > Enter [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/13/25 19:58:38.497 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/13/25 19:58:38.497 < Exit [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/13/25 19:58:44.403 (5.906s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:58:44.403 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:58:44.403 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-cvi --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-cvi\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/13/25 20:15:26.542 + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:58:44.403 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:58:44.403 (0s) > Enter [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/13/25 19:58:44.403 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/13/25 19:58:44.403 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-cvi --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-cvi\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/13/25 20:15:26.542 < Exit [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/13/25 20:15:26.542 (16m42.11s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 20:15:26.542 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 20:15:35.505 (8.963s) > Enter [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/13/25 20:15:35.505 STEP: Delete manifests - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:50 @ 10/13/25 20:15:35.505 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 20:15:35.506 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 20:15:37.72 (2.215s) < Exit [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/13/25 20:15:37.72 (2.215s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/13/25 20:15:37.721 + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.721 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 20:15:37.721 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.721 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.721 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 20:15:37.721 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.721 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.721 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 20:15:37.722 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.722 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.722 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 20:15:37.722 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.722 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.722 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 20:15:37.722 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.722 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.722 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 20:15:37.722 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.722 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.723 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 20:15:37.723 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.723 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.723 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 20:15:37.723 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.723 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.723 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 20:15:37.723 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 20:15:37.723 (0s) + + + > Enter [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/13/25 20:15:37.723 < Exit [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/13/25 20:15:39.727 (2.004s) > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 20:15:39.727 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 20:15:39.727 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/13/25 20:15:39.727 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/13/25 20:15:45.782 (6.055s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 20:15:45.782 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 20:15:45.782 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-always-on-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-always-on-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 @ 10/13/25 20:32:27.433 + > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 20:15:45.782 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 20:15:45.782 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/13/25 20:15:45.782 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/13/25 20:15:45.782 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/13/25 20:32:27.433 (16m41.657s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-always-on-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-always-on-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 @ 10/13/25 20:32:27.433 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/13/25 20:32:27.433 (16m41.657s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 20:32:27.433 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 20:32:38.217 (10.784s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:131 @ 10/13/25 20:32:38.218 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:177 @ 10/13/25 20:32:38.218 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:341 @ 10/13/25 20:32:38.218 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:367 @ 10/13/25 20:32:38.218 + + + + > Enter [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/13/25 20:32:38.218 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:51 @ 10/13/25 20:32:38.218 < Exit [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/13/25 20:32:38.218 (0s) > Enter [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/13/25 20:32:38.218 < Exit [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/13/25 20:32:38.218 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:82 @ 10/13/25 20:32:38.218 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:94 @ 10/13/25 20:32:38.218 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:106 @ 10/13/25 20:32:38.218 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:116 @ 10/13/25 20:32:38.218 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:132 @ 10/13/25 20:32:38.218 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:151 @ 10/13/25 20:32:38.219 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:157 @ 10/13/25 20:32:38.219 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:189 @ 10/13/25 20:32:38.219 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:216 @ 10/13/25 20:32:38.219 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:252 @ 10/13/25 20:32:38.219 + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/13/25 20:32:40.172 + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/13/25 20:32:38.219 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/13/25 20:32:40.172 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/13/25 20:32:40.172 (1.953s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/13/25 20:32:40.172 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/13/25 20:32:46.899 (6.727s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:100 @ 10/13/25 20:32:46.899 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:111 @ 10/13/25 20:32:46.9 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:122 @ 10/13/25 20:32:46.9 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:133 @ 10/13/25 20:32:46.9 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:144 @ 10/13/25 20:32:46.9 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:164 @ 10/13/25 20:32:46.9 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:185 @ 10/13/25 20:32:46.9 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:229 @ 10/13/25 20:32:46.9 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:290 @ 10/13/25 20:32:46.9 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:341 @ 10/13/25 20:32:46.9 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:367 @ 10/13/25 20:32:46.9 + + + + > Enter [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/13/25 20:32:46.9 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:41 @ 10/13/25 20:32:46.9 < Exit [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/13/25 20:32:46.901 (0s) > Enter [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/13/25 20:32:46.901 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 20:32:46.901 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 20:32:50.974 (4.073s) < Exit [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/13/25 20:32:50.974 (4.074s) + + + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/13/25 20:32:50.974 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/13/25 20:32:52.911 (1.936s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/13/25 20:32:52.911 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/13/25 20:32:57.148 (4.238s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 20:32:57.148 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 20:32:57.148 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/13/25 20:32:57.149 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:97 @ 10/13/25 20:32:57.149 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/13/25 20:33:13.355 (16.206s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 20:33:13.355 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 20:33:13.355 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-b --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-a --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 @ 10/13/25 20:49:55.399 + > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/13/25 20:33:13.355 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 @ 10/13/25 20:33:13.355 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-b --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-a --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 @ 10/13/25 20:49:55.399 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/13/25 20:49:55.4 (16m42.082s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 20:49:55.4 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 20:50:06.642 (11.243s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:118 @ 10/13/25 20:50:06.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:129 @ 10/13/25 20:50:06.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:142 @ 10/13/25 20:50:06.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:166 @ 10/13/25 20:50:06.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:174 @ 10/13/25 20:50:06.643 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:179 @ 10/13/25 20:50:06.643 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:187 @ 10/13/25 20:50:06.643 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:198 @ 10/13/25 20:50:06.643 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:209 @ 10/13/25 20:50:06.643 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:222 @ 10/13/25 20:50:06.643 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:228 @ 10/13/25 20:50:06.643 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:240 @ 10/13/25 20:50:06.643 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:250 @ 10/13/25 20:50:06.643 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:258 @ 10/13/25 20:50:06.643 + + + + > Enter [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/13/25 20:50:06.643 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:73 @ 10/13/25 20:50:06.643 < Exit [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/13/25 20:50:06.643 (0s) > Enter [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/13/25 20:50:06.643 < Exit [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/13/25 20:50:06.643 (0s) > Enter [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/13/25 20:50:06.643 < Exit [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/13/25 20:50:06.643 (0s) + + + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/13/25 20:50:06.643 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/13/25 20:50:08.611 (1.968s) > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 20:50:08.611 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 20:50:08.611 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/13/25 20:50:08.611 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/13/25 20:50:11.824 (3.213s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 20:50:11.824 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 20:50:11.824 (0s) + + + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 20:50:11.824 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 20:50:11.824 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/13/25 20:50:11.824 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:88 @ 10/13/25 20:50:11.824 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/13/25 20:50:16.909 (5.085s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 20:50:16.909 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 20:50:16.909 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-with-hotplug-standalone --namespace head-5073ae15-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/13/25 21:06:59.502 + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 20:50:16.909 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 20:50:16.909 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/13/25 20:50:16.909 STEP: VDs with consumers should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:99 @ 10/13/25 20:50:16.909 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-with-hotplug-standalone --namespace head-5073ae15-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/13/25 21:06:59.502 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/13/25 21:06:59.503 (16m42.6s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 21:06:59.503 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-disk-attachment -o yaml -l 'testcase=vm-disk-attachment' --show-managed-fields=true error: exit status 1 stderr: E1013 21:07:00.210993 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:00.325287 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:00.441951 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:00.558220 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:00.677913 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:00.791602 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:00.908362 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:01.020989 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:01.139392 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:01.255485 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:01.377778 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:01.494280 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:01.614721 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:01.731497 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:01.849735 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:01.963750 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:01.971753 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:01.973103 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:01.974411 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:01.975705 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:01.976966 50983 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-disk-attachment -o json -l 'testcase=vm-disk-attachment' stderr: E1013 21:07:02.918588 50984 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:03.031157 50984 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:03.148254 50984 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:03.260773 50984 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:03.382106 50984 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:03.502156 50984 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 21:07:03.624 (4.122s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/13/25 21:07:03.625 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/13/25 21:07:03.625 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/13/25 21:07:03.625 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/13/25 21:07:03.625 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/13/25 21:07:03.625 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/13/25 21:07:03.625 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/13/25 21:07:03.625 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/13/25 21:07:03.625 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/13/25 21:07:03.625 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/13/25 21:07:03.625 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-restore-safe Unexpected error: <*exec.ExitError | 0x14000708fc0>: exit status 1 { ProcessState: { pid: 50987, status: 256, rusage: { Utime: { Sec: 0, Usec: 38705, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 15633, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47415296, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3698, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 35, Nvcsw: 29, Nivcsw: 543, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:58 @ 10/13/25 21:07:04.286 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/13/25 21:07:03.626 [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-restore-safe Unexpected error: <*exec.ExitError | 0x14000708fc0>: exit status 1 { ProcessState: { pid: 50987, status: 256, rusage: { Utime: { Sec: 0, Usec: 38705, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 15633, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47415296, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3698, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 35, Nvcsw: 29, Nivcsw: 543, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:58 @ 10/13/25 21:07:04.286 < Exit [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/13/25 21:07:04.287 (661ms) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 21:07:04.287 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-restore-safe -o yaml -l 'testcase=vm-restore-safe' --show-managed-fields=true error: exit status 1 stderr: E1013 21:07:05.123032 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.237595 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.355162 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.469332 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.589557 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.703679 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.826924 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.942611 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.948036 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.949450 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.955757 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.957107 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.960964 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.962318 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.966446 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.967867 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.970903 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.972214 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.973526 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.974792 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:05.976056 50988 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-restore-safe -o json -l 'testcase=vm-restore-safe' stderr: E1013 21:07:06.511579 50989 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:06.632964 50989 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:06.767449 50989 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:06.884067 50989 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:07.002556 50989 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:07.118508 50989 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get pods) The list of pods is empty; nothing to dump. [PANICKED] Test Panicked In [AfterEach] at: /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/panic.go:262 @ 10/13/25 21:07:07.242 runtime error: invalid memory address or nil pointer dereference Full Stack Trace github.com/deckhouse/virtualization/tests/e2e.init.func23.3() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:70 +0xe4 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 21:07:07.242 (2.955s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/13/25 21:07:07.242 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:115 @ 10/13/25 21:07:07.242 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:158 @ 10/13/25 21:07:07.242 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:348 @ 10/13/25 21:07:07.243 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:374 @ 10/13/25 21:07:07.243 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-configuration Unexpected error: <*exec.ExitError | 0x14000694820>: exit status 1 { ProcessState: { pid: 50990, status: 256, rusage: { Utime: { Sec: 0, Usec: 41567, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 17213, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47906816, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3729, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 13, Nsignals: 77, Nvcsw: 96, Nivcsw: 645, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:55 @ 10/13/25 21:07:08.204 + > Enter [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/13/25 21:07:07.243 [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-configuration Unexpected error: <*exec.ExitError | 0x14000694820>: exit status 1 { ProcessState: { pid: 50990, status: 256, rusage: { Utime: { Sec: 0, Usec: 41567, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 17213, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47906816, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3729, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 13, Nsignals: 77, Nvcsw: 96, Nivcsw: 645, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:55 @ 10/13/25 21:07:08.204 < Exit [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/13/25 21:07:08.204 (962ms) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 21:07:08.204 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-configuration -o yaml -l 'testcase=vm-configuration' --show-managed-fields=true error: exit status 1 stderr: E1013 21:07:08.733758 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:08.850200 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:08.971421 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:08.973009 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:08.978224 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:08.979669 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:08.982727 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:08.984370 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:08.987410 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:08.988644 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:08.991907 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:08.993164 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:08.996081 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:08.997286 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:09.112690 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:09.227525 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:09.347053 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:09.463113 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:09.577387 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:09.690795 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:09.804474 50992 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-configuration -o json -l 'testcase=vm-configuration' stderr: E1013 21:07:10.647647 50993 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:10.760682 50993 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:10.893947 50993 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:11.006164 50993 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:11.124780 50993 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:11.238458 50993 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 21:07:11.359 (3.155s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/13/25 21:07:11.359 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/13/25 21:07:11.359 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:109 @ 10/13/25 21:07:11.36 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:123 @ 10/13/25 21:07:11.36 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:147 @ 10/13/25 21:07:11.36 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:161 @ 10/13/25 21:07:11.36 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:183 @ 10/13/25 21:07:11.36 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:202 @ 10/13/25 21:07:11.36 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:226 @ 10/13/25 21:07:11.36 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:240 @ 10/13/25 21:07:11.36 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:250 @ 10/13/25 21:07:11.36 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:265 @ 10/13/25 21:07:11.361 + + + + > Enter [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/13/25 21:07:11.361 [SKIPPED] Module SDN is disabled. Skipping all tests for module SDN. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:50 @ 10/13/25 21:07:12.149 < Exit [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/13/25 21:07:12.149 (788ms) > Enter [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/13/25 21:07:12.149 < Exit [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/13/25 21:07:12.149 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/13/25 21:07:12.149 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/13/25 21:07:12.15 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:111 @ 10/13/25 21:07:12.15 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:125 @ 10/13/25 21:07:12.15 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:141 @ 10/13/25 21:07:12.15 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:156 @ 10/13/25 21:07:12.15 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:169 @ 10/13/25 21:07:12.15 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-complex-test Unexpected error: <*exec.ExitError | 0x14000483bc0>: exit status 1 { ProcessState: { pid: 50994, status: 256, rusage: { Utime: { Sec: 0, Usec: 37338, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 21710, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47611904, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3718, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 40, Nvcsw: 30, Nivcsw: 568, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:58 @ 10/13/25 21:07:12.821 + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/13/25 21:07:12.151 [FAILED] kubectl create namespace head-5073ae15-end-to-end-complex-test Unexpected error: <*exec.ExitError | 0x14000483bc0>: exit status 1 { ProcessState: { pid: 50994, status: 256, rusage: { Utime: { Sec: 0, Usec: 37338, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 21710, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47611904, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3718, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 40, Nvcsw: 30, Nivcsw: 568, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:58 @ 10/13/25 21:07:12.821 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/13/25 21:07:12.821 (671ms) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 21:07:12.821 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-complex-test -o yaml -l 'testcase=complex-test' --show-managed-fields=true error: exit status 1 stderr: E1013 21:07:13.669984 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:13.796579 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:13.920367 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:14.041555 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:14.162526 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:14.279396 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:14.399756 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:14.528532 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:14.663405 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:14.778932 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:14.900460 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:15.018582 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:15.139687 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:15.256493 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:15.378412 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:15.496472 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:15.619726 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:15.736237 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:15.852246 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:15.969395 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:15.970939 50995 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-complex-test -o json -l 'testcase=complex-test' stderr: E1013 21:07:16.526189 50996 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:16.642312 50996 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:16.765328 50996 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:16.887087 50996 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:17.008699 50996 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:17.124043 50996 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 21:07:17.248 (4.426s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/13/25 21:07:17.248 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/13/25 21:07:17.248 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/13/25 21:07:17.248 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/13/25 21:07:17.249 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/13/25 21:07:17.249 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:141 @ 10/13/25 21:07:17.249 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:152 @ 10/13/25 21:07:17.249 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:162 @ 10/13/25 21:07:17.249 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:173 @ 10/13/25 21:07:17.249 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:184 @ 10/13/25 21:07:17.249 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:196 @ 10/13/25 21:07:17.249 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:220 @ 10/13/25 21:07:17.249 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:245 @ 10/13/25 21:07:17.249 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:268 @ 10/13/25 21:07:17.249 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:280 @ 10/13/25 21:07:17.249 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:299 @ 10/13/25 21:07:17.25 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:316 @ 10/13/25 21:07:17.25 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:341 @ 10/13/25 21:07:17.25 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:354 @ 10/13/25 21:07:17.25 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:358 @ 10/13/25 21:07:17.25 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:375 @ 10/13/25 21:07:17.25 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:388 @ 10/13/25 21:07:17.25 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:405 @ 10/13/25 21:07:17.25 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:444 @ 10/13/25 21:07:17.25 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:479 @ 10/13/25 21:07:17.25 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:497 @ 10/13/25 21:07:17.25 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:512 @ 10/13/25 21:07:17.25 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:528 @ 10/13/25 21:07:17.25 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:545 @ 10/13/25 21:07:17.25 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-sizing-policy Unexpected error: <*exec.ExitError | 0x140002920a0>: exit status 1 { ProcessState: { pid: 50997, status: 256, rusage: { Utime: { Sec: 0, Usec: 36558, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 18078, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47333376, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3698, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 46, Nvcsw: 39, Nivcsw: 601, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:64 @ 10/13/25 21:07:17.78 + > Enter [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/13/25 21:07:17.251 [FAILED] kubectl create namespace head-5073ae15-end-to-end-sizing-policy Unexpected error: <*exec.ExitError | 0x140002920a0>: exit status 1 { ProcessState: { pid: 50997, status: 256, rusage: { Utime: { Sec: 0, Usec: 36558, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 18078, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47333376, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3698, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 46, Nvcsw: 39, Nivcsw: 601, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:64 @ 10/13/25 21:07:17.78 < Exit [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/13/25 21:07:17.78 (530ms) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 21:07:17.78 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-sizing-policy -o yaml -l 'testcase=sizing-policy' --show-managed-fields=true error: exit status 1 stderr: E1013 21:07:18.284028 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:18.397181 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:18.518271 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:18.631633 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:18.749605 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:18.863334 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:18.989781 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:18.991266 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:19.000028 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:19.120781 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:19.239977 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:19.353830 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:19.471249 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:19.591269 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:19.713192 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:19.827543 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:19.945651 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:19.947123 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:19.948466 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:19.949837 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:19.951169 50999 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-sizing-policy -o json -l 'testcase=sizing-policy' stderr: E1013 21:07:20.588660 51000 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:20.704446 51000 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:20.824969 51000 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:20.941873 51000 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:20.946932 51000 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" E1013 21:07:20.948332 51000 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding" Error from server (InternalError): an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 21:07:21.07 (3.29s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/13/25 21:07:21.07 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/13/25 21:07:21.07 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/13/25 21:07:21.071 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:126 @ 10/13/25 21:07:21.071 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:140 @ 10/13/25 21:07:21.071 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:152 @ 10/13/25 21:07:21.071 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:157 @ 10/13/25 21:07:21.071 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:163 @ 10/13/25 21:07:21.071 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:176 @ 10/13/25 21:07:21.071 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:181 @ 10/13/25 21:07:21.071 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:187 @ 10/13/25 21:07:21.071 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:202 @ 10/13/25 21:07:21.071 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:216 @ 10/13/25 21:07:21.071 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:240 @ 10/13/25 21:07:21.071 + + + [FAILED] Expected success, but got an error: <*errors.StatusError | 0x14000955900>: an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get pods) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding (get pods)", Reason: "InternalError", Details: { Name: "", Group: "", Kind: "pods", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 502, }, } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/13/25 21:07:21.184 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 21:07:21.071 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 21:07:21.071 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 21:07:21.071 [FAILED] Expected success, but got an error: <*errors.StatusError | 0x14000955900>: an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get pods) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding (get pods)", Reason: "InternalError", Details: { Name: "", Group: "", Kind: "pods", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 502, }, } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/13/25 21:07:21.184 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 21:07:21.185 (113ms) + + + [FAILED] Expected success, but got an error: <*errors.StatusError | 0x14000e16000>: an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get projects.deckhouse.io) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding (get projects.deckhouse.io)", Reason: "InternalError", Details: { Name: "", Group: "deckhouse.io", Kind: "projects", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 502, }, } In [DeferCleanup (Suite)] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:177 @ 10/13/25 21:07:21.298 + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/13/25 21:07:21.185 [FAILED] Expected success, but got an error: <*errors.StatusError | 0x14000e16000>: an error on the server ("<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>") has prevented the request from succeeding (get projects.deckhouse.io) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding (get projects.deckhouse.io)", Reason: "InternalError", Details: { Name: "", Group: "deckhouse.io", Kind: "projects", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>502 Bad Gateway</title></head>\r\n<body>\r\n<center><h1>502 Bad Gateway</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 502, }, } In [DeferCleanup (Suite)] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:177 @ 10/13/25 21:07:21.298 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/13/25 21:07:21.298 (113ms) + + + \ No newline at end of file diff --git a/artifacts/clean-sds-20251012-114926-17692/junit.xml b/artifacts/clean-sds-20251012-114926-17692/junit.xml new file mode 100644 index 0000000000..017585cc00 --- /dev/null +++ b/artifacts/clean-sds-20251012-114926-17692/junit.xml @@ -0,0 +1,852 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 12:11:45.352 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 12:11:50.295 (4.943s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 12:11:50.295 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 12:11:50.295 (0s) + + + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/12/25 12:11:50.297 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/12/25 12:11:52.272 (1.975s) > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 12:11:52.272 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 12:11:52.272 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/12/25 12:11:52.273 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/12/25 12:11:56.037 (3.765s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 12:11:56.037 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 12:11:56.037 (0s) + + + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 12:11:56.038 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 12:11:56.038 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/12/25 12:11:56.038 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:88 @ 10/12/25 12:11:56.038 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/12/25 12:12:23.122 (27.084s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 12:12:23.122 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 12:12:23.122 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-with-hotplug-standalone --namespace head-5073ae15-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/12/25 12:29:05.243 + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 12:12:23.122 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 12:12:23.122 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/12/25 12:12:23.122 STEP: VDs with consumers should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:99 @ 10/12/25 12:12:23.122 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-with-hotplug-standalone --namespace head-5073ae15-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/12/25 12:29:05.243 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/12/25 12:29:05.244 (16m42.136s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 12:29:05.245 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 12:29:14.1 (8.855s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/12/25 12:29:14.1 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/12/25 12:29:14.1 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/12/25 12:29:14.1 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/12/25 12:29:14.1 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/12/25 12:29:14.1 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/12/25 12:29:14.101 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/12/25 12:29:14.101 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/12/25 12:29:14.101 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/12/25 12:29:14.101 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/12/25 12:29:14.101 + + + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/12/25 12:29:14.101 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/12/25 12:29:16.087 (1.986s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:62 @ 10/12/25 12:29:16.087 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:62 @ 10/12/25 12:29:43.311 (27.224s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 12:29:43.311 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 12:29:43.311 (0s) + + + > Enter [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/12/25 12:29:43.312 < Exit [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/12/25 12:29:44.472 (1.16s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 12:29:44.472 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 12:29:44.473 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/12/25 12:29:44.473 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:102 @ 10/12/25 12:29:44.473 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/12/25 12:29:50.499 (6.026s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 12:29:50.499 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 12:29:50.499 (0s) + + + > Enter [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/12/25 12:29:50.499 STEP: CVIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:113 @ 10/12/25 12:29:50.499 < Exit [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/12/25 12:29:53.452 (2.953s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 12:29:53.452 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 12:29:53.452 (0s) + + + > Enter [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/12/25 12:29:53.452 STEP: VMClasses should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:124 @ 10/12/25 12:29:53.452 < Exit [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/12/25 12:29:56.156 (2.704s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 12:29:56.156 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 12:29:56.156 (0s) + + + [FAILED] Timed out after 344.808s. Expected success, but got an error: <*errors.errorString | 0x14000c2c950>: cannot patch VMIP "head-5073ae15-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip { s: "cannot patch VMIP \"head-5073ae15-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:138 @ 10/12/25 12:35:40.99 + > Enter [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/12/25 12:29:56.156 [FAILED] Timed out after 344.808s. Expected success, but got an error: <*errors.errorString | 0x14000c2c950>: cannot patch VMIP "head-5073ae15-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip { s: "cannot patch VMIP \"head-5073ae15-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:138 @ 10/12/25 12:35:40.99 < Exit [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/12/25 12:35:40.991 (5m44.81s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 12:35:40.991 The list of pods is empty; nothing to dump. < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 12:35:47.49 (6.499s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:141 @ 10/12/25 12:35:47.49 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:152 @ 10/12/25 12:35:47.49 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:162 @ 10/12/25 12:35:47.49 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:173 @ 10/12/25 12:35:47.49 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:184 @ 10/12/25 12:35:47.49 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:196 @ 10/12/25 12:35:47.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:220 @ 10/12/25 12:35:47.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:245 @ 10/12/25 12:35:47.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:268 @ 10/12/25 12:35:47.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:280 @ 10/12/25 12:35:47.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:299 @ 10/12/25 12:35:47.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:316 @ 10/12/25 12:35:47.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:341 @ 10/12/25 12:35:47.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:354 @ 10/12/25 12:35:47.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:358 @ 10/12/25 12:35:47.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:375 @ 10/12/25 12:35:47.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:388 @ 10/12/25 12:35:47.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:405 @ 10/12/25 12:35:47.492 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:444 @ 10/12/25 12:35:47.492 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:479 @ 10/12/25 12:35:47.492 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:497 @ 10/12/25 12:35:47.492 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:512 @ 10/12/25 12:35:47.492 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:528 @ 10/12/25 12:35:47.492 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:545 @ 10/12/25 12:35:47.492 + + + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/12/25 12:35:47.492 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/12/25 12:35:49.493 (2.001s) > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 12:35:49.493 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 12:35:49.493 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/12/25 12:35:49.493 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/12/25 12:35:52.394 (2.9s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 12:35:52.394 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 12:35:52.394 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 12:35:52.394 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 12:35:52.394 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/12/25 12:35:52.394 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:75 @ 10/12/25 12:35:52.394 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/12/25 12:35:56.877 (4.483s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 12:35:56.877 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 12:35:56.877 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-label-annotation --namespace head-5073ae15-end-to-end-vm-label-annotation --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-label-annotation\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 @ 10/12/25 12:52:39.067 + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 12:35:56.877 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 12:35:56.877 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/12/25 12:35:56.877 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 @ 10/12/25 12:35:56.878 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-label-annotation --namespace head-5073ae15-end-to-end-vm-label-annotation --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-label-annotation\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 @ 10/12/25 12:52:39.067 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/12/25 12:52:39.068 (16m42.174s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 12:52:39.068 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 12:52:45.297 (6.229s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/12/25 12:52:45.297 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/12/25 12:52:45.297 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/12/25 12:52:45.297 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/12/25 12:52:45.297 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/12/25 12:52:45.297 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/12/25 12:52:45.297 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/12/25 12:52:45.298 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/12/25 12:52:45.298 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/12/25 12:52:45.298 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/12/25 12:52:45.298 + + + + > Enter [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/12/25 12:52:45.298 [SKIPPED] Module SDN is disabled. Skipping all tests for module SDN. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:50 @ 10/12/25 12:52:45.972 < Exit [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/12/25 12:52:45.972 (674ms) > Enter [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/12/25 12:52:45.972 < Exit [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/12/25 12:52:45.972 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/12/25 12:52:45.973 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/12/25 12:52:45.973 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:111 @ 10/12/25 12:52:45.973 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:125 @ 10/12/25 12:52:45.973 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:141 @ 10/12/25 12:52:45.973 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:156 @ 10/12/25 12:52:45.973 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:169 @ 10/12/25 12:52:45.973 + + + + > Enter [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/12/25 12:52:45.974 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:51 @ 10/12/25 12:52:45.974 < Exit [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/12/25 12:52:45.974 (0s) > Enter [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/12/25 12:52:45.974 < Exit [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/12/25 12:52:45.974 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:82 @ 10/12/25 12:52:45.974 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:94 @ 10/12/25 12:52:45.974 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:106 @ 10/12/25 12:52:45.974 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:116 @ 10/12/25 12:52:45.974 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:132 @ 10/12/25 12:52:45.974 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:151 @ 10/12/25 12:52:45.975 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:157 @ 10/12/25 12:52:45.975 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:189 @ 10/12/25 12:52:45.975 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:216 @ 10/12/25 12:52:45.975 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:252 @ 10/12/25 12:52:45.975 + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.975 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 12:52:45.975 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.975 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.976 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 12:52:45.976 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.976 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.976 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 12:52:45.976 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.976 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.976 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 12:52:45.976 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.976 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.976 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 12:52:45.976 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.976 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.976 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 12:52:45.977 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.977 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.977 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 12:52:45.977 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.977 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.977 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 12:52:45.977 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.977 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.977 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 12:52:45.977 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 12:52:45.977 (0s) + + + > Enter [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/12/25 12:52:45.977 < Exit [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/12/25 12:52:47.949 (1.971s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/12/25 12:52:47.949 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/12/25 12:52:57.06 (9.111s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/12/25 12:52:57.06 < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/12/25 12:52:57.06 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:6, cap:6>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-c-and-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-c-and-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-selector --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-selector\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-d --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-d\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-affinity --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-affinity\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-b-not-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-b-not-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-a-not-b --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-a-not-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:103 @ 10/12/25 13:09:47.23 + > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/12/25 12:52:57.06 STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/12/25 12:52:57.06 END STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/12/25 12:53:02.278 (5.218s) STEP: `VirtualMachineClasses` should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:95 @ 10/12/25 12:53:02.278 END STEP: `VirtualMachineClasses` should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:95 @ 10/12/25 12:53:05.02 (2.742s) STEP: `VirtualDisks` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:102 @ 10/12/25 12:53:05.02 END STEP: `VirtualDisks` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:102 @ 10/12/25 13:09:47.23 (16m42.194s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:6, cap:6>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-c-and-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-c-and-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-selector --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-selector\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-d --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-d\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-affinity --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-affinity\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-b-not-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-b-not-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-a-not-b --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-a-not-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:103 @ 10/12/25 13:09:47.23 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/12/25 13:09:47.231 (16m50.155s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/12/25 13:09:47.231 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/12/25 13:09:53.537 (6.306s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:120 @ 10/12/25 13:09:53.537 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:283 @ 10/12/25 13:09:53.537 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:372 @ 10/12/25 13:09:53.537 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:467 @ 10/12/25 13:09:53.537 + + + > Enter [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/12/25 13:09:53.538 < Exit [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/12/25 13:09:56.749 (3.211s) > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/12/25 13:09:56.749 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/12/25 13:09:56.749 (0s) > Enter [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/12/25 13:09:56.749 STEP: Create a vmip automatically and check its binding with a lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:81 @ 10/12/25 13:09:56.749 STEP: Remove label from the lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:87 @ 10/12/25 13:09:57.702 STEP: Wait for the label to be restored by the controller - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:96 @ 10/12/25 13:09:57.82 < Exit [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/12/25 13:09:58.061 (1.312s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/12/25 13:09:58.061 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/12/25 13:09:58.061 (0s) + + + > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/12/25 13:09:58.061 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/12/25 13:09:58.061 (0s) > Enter [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/12/25 13:09:58.061 STEP: Create an intermediate vmip automatically to allocate a new ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:109 @ 10/12/25 13:09:58.062 STEP: Delete the intermediate vmip automatically and check that the lease is released - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:115 @ 10/12/25 13:09:58.435 STEP: Reuse the released lease with a static vmip - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:124 @ 10/12/25 13:09:58.665 STEP: Delete the static vmip and lease, then create another static vmip with this ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:132 @ 10/12/25 13:09:59.119 < Exit [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/12/25 13:09:59.92 (1.858s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/12/25 13:09:59.92 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/12/25 13:09:59.92 (0s) + + + > Enter [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/12/25 13:09:59.92 < Exit [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/12/25 13:10:01.86 (1.939s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/12/25 13:10:01.86 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/12/25 13:10:07.661 (5.802s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/12/25 13:10:07.662 < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/12/25 13:10:07.662 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:77 @ 10/12/25 13:26:49.778 + > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/12/25 13:10:07.662 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:76 @ 10/12/25 13:10:07.662 [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:77 @ 10/12/25 13:26:49.778 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/12/25 13:26:49.78 (16m42.134s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/12/25 13:26:49.78 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/12/25 13:26:55.887 (6.107s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:86 @ 10/12/25 13:26:55.887 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:100 @ 10/12/25 13:26:55.887 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:116 @ 10/12/25 13:26:55.888 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:131 @ 10/12/25 13:26:55.888 + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/12/25 13:26:57.898 + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/12/25 13:26:55.888 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/12/25 13:26:57.898 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/12/25 13:26:57.898 (2.01s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/12/25 13:26:57.898 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/12/25 13:27:04.013 (6.115s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:100 @ 10/12/25 13:27:04.014 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:111 @ 10/12/25 13:27:04.014 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:122 @ 10/12/25 13:27:04.014 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:133 @ 10/12/25 13:27:04.014 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:144 @ 10/12/25 13:27:04.014 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:164 @ 10/12/25 13:27:04.014 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:185 @ 10/12/25 13:27:04.014 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:229 @ 10/12/25 13:27:04.014 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:290 @ 10/12/25 13:27:04.014 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:341 @ 10/12/25 13:27:04.015 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:367 @ 10/12/25 13:27:04.015 + + + > Enter [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/12/25 13:27:04.015 < Exit [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/12/25 13:27:06.005 (1.991s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/12/25 13:27:06.005 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/12/25 13:27:10.991 (4.986s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/12/25 13:27:10.991 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/12/25 13:27:10.991 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/12/25 13:27:10.992 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:89 @ 10/12/25 13:27:10.992 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/12/25 13:27:16.139 (5.147s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/12/25 13:27:16.139 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/12/25 13:27:16.139 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:4>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-manual-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-manual-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:100 @ 10/12/25 13:43:58.258 + > Enter [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/12/25 13:27:16.14 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:4>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-manual-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-manual-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:100 @ 10/12/25 13:43:58.258 < Exit [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/12/25 13:43:58.259 (16m42.134s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/12/25 13:43:58.259 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/12/25 13:44:04.413 (6.153s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:109 @ 10/12/25 13:44:04.413 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:123 @ 10/12/25 13:44:04.413 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:147 @ 10/12/25 13:44:04.413 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:161 @ 10/12/25 13:44:04.413 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:183 @ 10/12/25 13:44:04.414 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:202 @ 10/12/25 13:44:04.414 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:226 @ 10/12/25 13:44:04.414 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:240 @ 10/12/25 13:44:04.414 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:250 @ 10/12/25 13:44:04.414 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:265 @ 10/12/25 13:44:04.414 + + + + > Enter [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/12/25 13:44:04.414 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:41 @ 10/12/25 13:44:04.414 < Exit [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/12/25 13:44:04.414 (0s) > Enter [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/12/25 13:44:04.414 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 13:44:04.414 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 13:44:09.143 (4.729s) < Exit [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/12/25 13:44:09.143 (4.729s) + + + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/12/25 13:44:09.144 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/12/25 13:44:11.095 (1.951s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/12/25 13:44:11.095 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/12/25 13:44:15.048 (3.952s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 13:44:15.048 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 13:44:15.048 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/12/25 13:44:15.048 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:97 @ 10/12/25 13:44:15.048 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/12/25 13:44:31.821 (16.773s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 13:44:31.821 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 13:44:31.821 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-a --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-b --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 @ 10/12/25 14:01:14.366 + > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/12/25 13:44:31.821 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 @ 10/12/25 13:44:31.821 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-a --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-b --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 @ 10/12/25 14:01:14.366 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/12/25 14:01:14.366 (16m42.606s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 14:01:14.366 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 14:01:20.225 (5.859s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:118 @ 10/12/25 14:01:20.225 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:129 @ 10/12/25 14:01:20.226 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:142 @ 10/12/25 14:01:20.226 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:166 @ 10/12/25 14:01:20.226 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:174 @ 10/12/25 14:01:20.226 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:179 @ 10/12/25 14:01:20.226 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:187 @ 10/12/25 14:01:20.226 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:198 @ 10/12/25 14:01:20.226 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:209 @ 10/12/25 14:01:20.226 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:222 @ 10/12/25 14:01:20.226 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:228 @ 10/12/25 14:01:20.226 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:240 @ 10/12/25 14:01:20.226 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:250 @ 10/12/25 14:01:20.226 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:258 @ 10/12/25 14:01:20.226 + + + > Enter [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/12/25 14:01:20.226 < Exit [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/12/25 14:01:20.227 (1ms) > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:01:20.227 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:01:20.227 (0s) > Enter [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/12/25 14:01:20.227 < Exit [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/12/25 14:01:21.78 (1.553s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:01:21.78 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:01:21.78 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:01:21.78 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:01:21.78 (0s) > Enter [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/12/25 14:01:21.78 STEP: Project should be deployed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:73 @ 10/12/25 14:01:21.78 < Exit [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/12/25 14:01:24.071 (2.291s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:01:24.071 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:01:24.071 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:01:24.071 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:01:24.071 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/12/25 14:01:24.071 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/12/25 14:01:26.65 (2.579s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:01:26.65 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:01:26.65 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:01:26.65 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:01:26.65 (0s) > Enter [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/12/25 14:01:26.65 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/12/25 14:01:26.651 < Exit [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/12/25 14:01:32.777 (6.126s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:01:32.777 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:01:32.777 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-cvi --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-cvi\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/12/25 14:18:14.872 + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:01:32.777 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 14:01:32.777 (0s) > Enter [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/12/25 14:01:32.777 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/12/25 14:01:32.777 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-cvi --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-cvi\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/12/25 14:18:14.872 < Exit [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/12/25 14:18:14.873 (16m42.109s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:18:14.873 The list of pods is empty; nothing to dump. < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 14:18:21.248 (6.374s) > Enter [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/12/25 14:18:21.248 STEP: Delete manifests - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:50 @ 10/12/25 14:18:21.248 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 14:18:21.248 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 14:18:24.091 (2.843s) < Exit [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/12/25 14:18:24.091 (2.843s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/12/25 14:18:24.091 + + + > Enter [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/12/25 14:18:24.091 < Exit [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/12/25 14:18:26.037 (1.946s) > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 14:18:26.037 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 14:18:26.038 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/12/25 14:18:26.038 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/12/25 14:18:27.639 (1.602s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 14:18:27.639 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 14:18:27.639 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root --namespace head-5073ae15-end-to-end-vm-versions --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:69 @ 10/12/25 14:35:09.826 + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 14:18:27.639 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 14:18:27.639 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/12/25 14:18:27.639 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:68 @ 10/12/25 14:18:27.639 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root --namespace head-5073ae15-end-to-end-vm-versions --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:69 @ 10/12/25 14:35:09.826 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/12/25 14:35:09.826 (16m42.139s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 14:35:09.827 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 14:35:18.608 (8.782s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/12/25 14:35:18.609 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/12/25 14:35:18.609 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/12/25 14:35:18.609 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/12/25 14:35:18.609 + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:51 @ 10/12/25 14:35:20.101 + > Enter [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/12/25 14:35:18.609 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:51 @ 10/12/25 14:35:20.101 < Exit [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/12/25 14:35:20.102 (1.492s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/12/25 14:35:20.102 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/12/25 14:35:25.641 (5.54s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:94 @ 10/12/25 14:35:25.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:103 @ 10/12/25 14:35:25.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:114 @ 10/12/25 14:35:25.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:128 @ 10/12/25 14:35:25.642 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:139 @ 10/12/25 14:35:25.642 + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.642 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/12/25 14:35:25.643 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.643 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.643 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/12/25 14:35:25.643 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.643 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.643 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/12/25 14:35:25.643 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.643 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.643 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/12/25 14:35:25.643 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.643 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.643 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/12/25 14:35:25.643 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.644 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.644 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/12/25 14:35:25.644 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.644 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.644 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/12/25 14:35:25.644 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.644 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.644 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/12/25 14:35:25.644 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.644 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.644 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/12/25 14:35:25.644 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.644 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.644 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/12/25 14:35:25.644 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.644 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.645 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/12/25 14:35:25.645 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.645 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.645 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/12/25 14:35:25.645 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/12/25 14:35:25.645 (0s) + + + + > Enter [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/12/25 14:35:25.645 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:73 @ 10/12/25 14:35:25.645 < Exit [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/12/25 14:35:25.645 (0s) > Enter [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/12/25 14:35:25.645 < Exit [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/12/25 14:35:25.645 (0s) > Enter [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/12/25 14:35:25.645 < Exit [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/12/25 14:35:25.645 (0s) + + + > Enter [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/12/25 14:35:25.645 < Exit [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/12/25 14:35:27.2 (1.555s) > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/12/25 14:35:27.201 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/12/25 14:35:27.201 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/12/25 14:35:27.201 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/12/25 14:35:31.38 (4.18s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/12/25 14:35:31.38 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/12/25 14:35:31.38 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-safe --namespace head-5073ae15-end-to-end-vm-restore-safe --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-safe\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:107 @ 10/12/25 14:52:13.015 + > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/12/25 14:35:31.38 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/12/25 14:35:31.38 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/12/25 14:35:31.38 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/12/25 14:35:31.381 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/12/25 14:52:13.015 (16m41.666s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-safe --namespace head-5073ae15-end-to-end-vm-restore-safe --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-safe\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:107 @ 10/12/25 14:52:13.015 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/12/25 14:52:13.016 (16m41.667s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/12/25 14:52:13.016 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/12/25 14:52:19.357 (6.341s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:115 @ 10/12/25 14:52:19.357 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:158 @ 10/12/25 14:52:19.357 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:348 @ 10/12/25 14:52:19.357 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:374 @ 10/12/25 14:52:19.357 + + + > Enter [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/12/25 14:52:19.357 < Exit [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/12/25 14:52:21.334 (1.976s) > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 14:52:21.334 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 14:52:21.334 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/12/25 14:52:21.334 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/12/25 14:52:27.671 (6.337s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 14:52:27.671 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 14:52:27.671 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 14:52:27.671 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 14:52:27.671 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/12/25 14:52:27.671 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:91 @ 10/12/25 14:52:27.671 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/12/25 14:52:31.658 (3.987s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 14:52:31.658 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 14:52:31.658 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=WaitForFirstConsumer'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-blank-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=WaitForFirstConsumer' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-blank-not-existing-vmclass-with-changing\n\nwaited for: 'jsonpath={.status.phase}=WaitForFirstConsumer'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=WaitForFirstConsumer' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-not-existing-vmclass-with-changing\n\nwaited for: 'jsonpath={.status.phase}=WaitForFirstConsumer'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:103 @ 10/12/25 15:09:13.852 + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 14:52:31.658 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 14:52:31.658 (0s) > Enter [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/12/25 14:52:31.658 STEP: VDs should be in WaitForFirstConsumer phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:102 @ 10/12/25 14:52:31.658 [FAILED] should observe resources in ''jsonpath={.status.phase}=WaitForFirstConsumer'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-blank-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=WaitForFirstConsumer' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-blank-not-existing-vmclass-with-changing\n\nwaited for: 'jsonpath={.status.phase}=WaitForFirstConsumer'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=WaitForFirstConsumer' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-not-existing-vmclass-with-changing\n\nwaited for: 'jsonpath={.status.phase}=WaitForFirstConsumer'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:103 @ 10/12/25 15:09:13.852 < Exit [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/12/25 15:09:13.853 (16m42.154s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 15:09:13.853 The list of pods is empty; nothing to dump. < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 15:09:20.089 (6.236s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/12/25 15:09:20.089 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:126 @ 10/12/25 15:09:20.089 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:140 @ 10/12/25 15:09:20.089 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:152 @ 10/12/25 15:09:20.089 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:157 @ 10/12/25 15:09:20.09 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:163 @ 10/12/25 15:09:20.09 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:176 @ 10/12/25 15:09:20.09 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:181 @ 10/12/25 15:09:20.09 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:187 @ 10/12/25 15:09:20.09 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:202 @ 10/12/25 15:09:20.09 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:216 @ 10/12/25 15:09:20.09 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:240 @ 10/12/25 15:09:20.094 + + + [FAILED] Warning: resource namespaces/head-5073ae15-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "ceph-pool-r2-csi-rbd" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x140007362e0>: exit status 1 { ProcessState: { pid: 72312, status: 256, rusage: { Utime: { Sec: 0, Usec: 292584, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 59199, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 69058560, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5029, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 48, Msgrcv: 53, Nsignals: 311, Nvcsw: 247, Nivcsw: 2685, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/12/25 15:09:26.912 + > Enter [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/12/25 15:09:20.094 < Exit [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/12/25 15:09:23.129 (3.035s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/12/25 15:09:23.129 [FAILED] Warning: resource namespaces/head-5073ae15-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "ceph-pool-r2-csi-rbd" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x140007362e0>: exit status 1 { ProcessState: { pid: 72312, status: 256, rusage: { Utime: { Sec: 0, Usec: 292584, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 59199, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 69058560, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5029, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 48, Msgrcv: 53, Nsignals: 311, Nvcsw: 247, Nivcsw: 2685, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/12/25 15:09:26.912 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/12/25 15:09:26.913 (3.783s) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/12/25 15:09:26.913 The list of pods is empty; nothing to dump. < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/12/25 15:09:32.488 (5.575s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:112 @ 10/12/25 15:09:32.488 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:147 @ 10/12/25 15:09:32.488 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:190 @ 10/12/25 15:09:32.489 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:196 @ 10/12/25 15:09:32.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:204 @ 10/12/25 15:09:32.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:224 @ 10/12/25 15:09:32.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:236 @ 10/12/25 15:09:32.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:258 @ 10/12/25 15:09:32.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:284 @ 10/12/25 15:09:32.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:294 @ 10/12/25 15:09:32.491 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:309 @ 10/12/25 15:09:32.491 + + + [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/12/25 15:11:37.807 This is the Progress Report generated when the suite timeout occurred: VirtualMachineEvacuation Evacuation (Spec Runtime: 2m5.316s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 In [It] (Node Runtime: 1m58.24s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 At [By Step] Virtual machine agents should be ready (Step Runtime: 1m58.24s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 Spec Goroutine goroutine 919 [sync.WaitGroup.Wait] sync.runtime_SemacquireWaitGroup(0x102928b30?, 0xa0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000a82010) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x1400041a0c0, 0x2, 0x1043a1e58?}, {0x1040bea2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x140005feb40, {0x1400039fe30, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1040bea2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x140005feb40, {0x1400039fe30, 0x26}, {0x14000976030, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1040bea2d, 0x2a}, {0x10407ba0a, 0x7}, {{0x0, 0x0, 0x0}, 0x140005feb40, {0x1400039fe30, 0x26}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x140005feb40, {0x1400039fe30, 0x26}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func17.5() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 | It("Evacuation", func() { | By("Virtual machine agents should be ready") > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x1400074cdc0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 24 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 260 [syscall] syscall.syscall6(0x100a82050?, 0x106ca5110?, 0x106c99d58?, 0x90?, 0x105f24480?, 0x1400022a360?, 0x14000f8ea58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000f8ea88?, 0x1029aa6fc?, 0x90?, 0x1049a6120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002c8150?, 0x14000f8eac4, 0x14000a82050?, 0x140002c80e0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000918140) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000826408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x1400072e000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x1400072e000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140003f19e0?, 0x104a3de88?, 0x140002c8000?}}, {0x104a5e488?, 0x140002c8000?}, {0x14000824000?, 0xc233058032aaf280?}, {0x104a36d80, 0x140009180c0}, {0x104a36d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140003f19e0?, 0x105f4b728?, 0x105f21fa0?}}, {0x104a5e488, 0x140002c8000}, {0x14000824000, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x104a62a38?, 0x140002cac00?}, {0x10407bfba?, 0x140009b7e88?}}, {0x1040bea2d, 0x2a}, {0x140002ea3a0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 919 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 259 [syscall] syscall.syscall6(0x10099c030?, 0x1323d3b50?, 0x106c98f30?, 0x90?, 0x14000101008?, 0x14000afa090?, 0x140000b4a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140000b4a88?, 0x1029aa6fc?, 0x90?, 0x1049a6120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002ec5b0?, 0x140000b4ac4, 0x1400099c030?, 0x140002ec3f0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000124200) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x1400091a408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000197680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000197680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140003f19e0?, 0x104a3de88?, 0x140002ec000?}}, {0x104a5e488?, 0x140002ec000?}, {0x14000510000?, 0xc233058032ab1990?}, {0x104a36d80, 0x14000124000}, {0x104a36d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140003f19e0?, 0x105f4b728?, 0x105f21fa0?}}, {0x104a5e488, 0x140002ec000}, {0x14000510000, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x104a62a38?, 0x140002cac00?}, {0x10407bfba?, 0x1400050c688?}}, {0x1040bea2d, 0x2a}, {0x140002ea380, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 919 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/12/25 15:09:32.491 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/12/25 15:09:34.464 (1.973s) > Enter [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/12/25 15:09:34.464 < Exit [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/12/25 15:09:39.567 (5.103s) > Enter [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/12/25 15:09:39.567 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 @ 10/12/25 15:09:39.567 [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/12/25 15:11:37.807 This is the Progress Report generated when the suite timeout occurred: VirtualMachineEvacuation Evacuation (Spec Runtime: 2m5.316s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 In [It] (Node Runtime: 1m58.24s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 At [By Step] Virtual machine agents should be ready (Step Runtime: 1m58.24s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 Spec Goroutine goroutine 919 [sync.WaitGroup.Wait] sync.runtime_SemacquireWaitGroup(0x102928b30?, 0xa0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000a82010) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x1400041a0c0, 0x2, 0x1043a1e58?}, {0x1040bea2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x140005feb40, {0x1400039fe30, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1040bea2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x140005feb40, {0x1400039fe30, 0x26}, {0x14000976030, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1040bea2d, 0x2a}, {0x10407ba0a, 0x7}, {{0x0, 0x0, 0x0}, 0x140005feb40, {0x1400039fe30, 0x26}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x140005feb40, {0x1400039fe30, 0x26}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func17.5() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 | It("Evacuation", func() { | By("Virtual machine agents should be ready") > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x1400074cdc0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 24 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 260 [syscall] syscall.syscall6(0x100a82050?, 0x106ca5110?, 0x106c99d58?, 0x90?, 0x105f24480?, 0x1400022a360?, 0x14000f8ea58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000f8ea88?, 0x1029aa6fc?, 0x90?, 0x1049a6120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002c8150?, 0x14000f8eac4, 0x14000a82050?, 0x140002c80e0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000918140) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000826408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x1400072e000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x1400072e000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140003f19e0?, 0x104a3de88?, 0x140002c8000?}}, {0x104a5e488?, 0x140002c8000?}, {0x14000824000?, 0xc233058032aaf280?}, {0x104a36d80, 0x140009180c0}, {0x104a36d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140003f19e0?, 0x105f4b728?, 0x105f21fa0?}}, {0x104a5e488, 0x140002c8000}, {0x14000824000, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x104a62a38?, 0x140002cac00?}, {0x10407bfba?, 0x140009b7e88?}}, {0x1040bea2d, 0x2a}, {0x140002ea3a0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 919 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 259 [syscall] syscall.syscall6(0x10099c030?, 0x1323d3b50?, 0x106c98f30?, 0x90?, 0x14000101008?, 0x14000afa090?, 0x140000b4a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140000b4a88?, 0x1029aa6fc?, 0x90?, 0x1049a6120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002ec5b0?, 0x140000b4ac4, 0x1400099c030?, 0x140002ec3f0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000124200) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x1400091a408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000197680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000197680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140003f19e0?, 0x104a3de88?, 0x140002ec000?}}, {0x104a5e488?, 0x140002ec000?}, {0x14000510000?, 0xc233058032ab1990?}, {0x104a36d80, 0x14000124000}, {0x104a36d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140003f19e0?, 0x105f4b728?, 0x105f21fa0?}}, {0x104a5e488, 0x140002ec000}, {0x14000510000, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x104a62a38?, 0x140002cac00?}, {0x10407bfba?, 0x1400050c688?}}, {0x1040bea2d, 0x2a}, {0x140002ea380, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 919 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/12/25 15:11:37.814 (1m58.248s) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/12/25 15:11:37.814 The list of pods is empty; nothing to dump. STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 15:11:43.937 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 15:11:48.402 (4.465s) < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/12/25 15:11:48.402 (10.588s) + + + + + + + + + + + + + + + + + + + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x140006a9368>: the container "virtualization-controller" was not found: virtualization-controller-75845f987f-9mg56 the container "virtualization-controller" was restarted: virtualization-controller-7b6c96b4c9-bprrz the container "virtualization-controller" was restarted: virtualization-controller-7b6c96b4c9-swkng { errs: [ <*errors.joinError | 0x140006a9350>{ errs: [ <*errors.joinError | 0x140006a9338>{ errs: [ <*errors.errorString | 0x140003f1d10>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-75845f987f-9mg56", }, ], }, <*errors.errorString | 0x140003f1d50>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-7b6c96b4c9-bprrz", }, ], }, <*errors.errorString | 0x140003f1d70>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-7b6c96b4c9-swkng", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/12/25 15:11:49.231 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/12/25 15:11:48.404 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/12/25 15:11:48.404 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/12/25 15:11:48.404 [FAILED] Expected success, but got an error: <*errors.joinError | 0x140006a9368>: the container "virtualization-controller" was not found: virtualization-controller-75845f987f-9mg56 the container "virtualization-controller" was restarted: virtualization-controller-7b6c96b4c9-bprrz the container "virtualization-controller" was restarted: virtualization-controller-7b6c96b4c9-swkng { errs: [ <*errors.joinError | 0x140006a9350>{ errs: [ <*errors.joinError | 0x140006a9338>{ errs: [ <*errors.errorString | 0x140003f1d10>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-75845f987f-9mg56", }, ], }, <*errors.errorString | 0x140003f1d50>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-7b6c96b4c9-bprrz", }, ], }, <*errors.errorString | 0x140003f1d70>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-7b6c96b4c9-swkng", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/12/25 15:11:49.231 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/12/25 15:11:49.231 (826ms) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/12/25 15:11:49.231 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/12/25 15:12:04.8 (15.569s) + + + \ No newline at end of file diff --git a/artifacts/clean-sds-20251013-081321-32554/junit.xml b/artifacts/clean-sds-20251013-081321-32554/junit.xml new file mode 100644 index 0000000000..711121dbf0 --- /dev/null +++ b/artifacts/clean-sds-20251013-081321-32554/junit.xml @@ -0,0 +1,857 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 08:32:13.545 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 08:32:18.312 (4.767s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 08:32:18.312 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 08:32:18.312 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.314 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 08:32:18.315 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.315 (1ms) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.315 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 08:32:18.315 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.315 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.315 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 08:32:18.315 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.315 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.315 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 08:32:18.315 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.315 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.316 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 08:32:18.316 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.316 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.316 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 08:32:18.316 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.316 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.316 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 08:32:18.316 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.316 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.316 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 08:32:18.316 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.316 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.316 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 08:32:18.316 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.316 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.316 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 08:32:18.316 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.316 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.317 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 08:32:18.317 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.317 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.317 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 08:32:18.317 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 08:32:18.317 (0s) + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:51 @ 10/13/25 08:32:19.811 + > Enter [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/13/25 08:32:18.317 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:51 @ 10/13/25 08:32:19.811 < Exit [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/13/25 08:32:19.811 (1.494s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/13/25 08:32:19.811 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/13/25 08:32:25.341 (5.531s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:94 @ 10/13/25 08:32:25.341 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:103 @ 10/13/25 08:32:25.342 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:114 @ 10/13/25 08:32:25.342 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:128 @ 10/13/25 08:32:25.342 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:139 @ 10/13/25 08:32:25.342 + + + > Enter [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/13/25 08:32:25.342 < Exit [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/13/25 08:32:26.838 (1.496s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/13/25 08:32:26.838 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/13/25 08:32:31.869 (5.031s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 08:32:31.869 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 08:32:31.869 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/13/25 08:32:31.869 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:89 @ 10/13/25 08:32:31.869 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/13/25 08:32:39.693 (7.824s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 08:32:39.693 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 08:32:39.693 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:4>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-manual-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-manual-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:100 @ 10/13/25 08:49:21.823 + > Enter [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/13/25 08:32:39.694 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:4>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-manual-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-manual-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:100 @ 10/13/25 08:49:21.823 < Exit [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/13/25 08:49:21.824 (16m42.136s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 08:49:21.824 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 08:49:32.7 (10.876s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:109 @ 10/13/25 08:49:32.7 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:123 @ 10/13/25 08:49:32.701 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:147 @ 10/13/25 08:49:32.701 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:161 @ 10/13/25 08:49:32.701 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:183 @ 10/13/25 08:49:32.701 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:202 @ 10/13/25 08:49:32.701 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:226 @ 10/13/25 08:49:32.701 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:240 @ 10/13/25 08:49:32.701 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:250 @ 10/13/25 08:49:32.701 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:265 @ 10/13/25 08:49:32.701 + + + > Enter [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/13/25 08:49:32.701 < Exit [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/13/25 08:49:36.273 (3.572s) > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 08:49:36.273 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 08:49:36.273 (0s) > Enter [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/13/25 08:49:36.273 STEP: Create a vmip automatically and check its binding with a lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:81 @ 10/13/25 08:49:36.273 STEP: Remove label from the lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:87 @ 10/13/25 08:49:36.668 STEP: Wait for the label to be restored by the controller - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:96 @ 10/13/25 08:49:36.781 < Exit [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/13/25 08:49:37.009 (735ms) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 08:49:37.009 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 08:49:37.009 (0s) + + + > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 08:49:37.009 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 08:49:37.009 (0s) > Enter [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/13/25 08:49:37.009 STEP: Create an intermediate vmip automatically to allocate a new ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:109 @ 10/13/25 08:49:37.009 STEP: Delete the intermediate vmip automatically and check that the lease is released - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:115 @ 10/13/25 08:49:37.351 STEP: Reuse the released lease with a static vmip - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:124 @ 10/13/25 08:49:37.576 STEP: Delete the static vmip and lease, then create another static vmip with this ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:132 @ 10/13/25 08:49:38.03 < Exit [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/13/25 08:49:38.786 (1.777s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 08:49:38.786 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 08:49:38.786 (0s) + + + > Enter [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/13/25 08:49:38.786 < Exit [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/13/25 08:49:40.779 (1.993s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/13/25 08:49:40.779 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/13/25 08:49:46.318 (5.539s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 08:49:46.318 < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 08:49:46.318 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:77 @ 10/13/25 09:06:28.457 + > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/13/25 08:49:46.319 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:76 @ 10/13/25 08:49:46.319 [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:77 @ 10/13/25 09:06:28.457 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/13/25 09:06:28.457 (16m42.122s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 09:06:28.457 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 09:06:34.682 (6.225s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:86 @ 10/13/25 09:06:34.683 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:100 @ 10/13/25 09:06:34.683 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:116 @ 10/13/25 09:06:34.683 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:131 @ 10/13/25 09:06:34.683 + + + + > Enter [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/13/25 09:06:34.683 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:41 @ 10/13/25 09:06:34.684 < Exit [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/13/25 09:06:34.684 (0s) > Enter [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/13/25 09:06:34.684 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 09:06:34.684 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 09:06:39.392 (4.709s) < Exit [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/13/25 09:06:39.392 (4.709s) + + + [FAILED] Warning: resource namespaces/head-5073ae15-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "ceph-pool-r2-csi-rbd" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x140001bc9e0>: exit status 1 { ProcessState: { pid: 11922, status: 256, rusage: { Utime: { Sec: 0, Usec: 297871, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 61792, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 68583424, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 4995, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 49, Msgrcv: 56, Nsignals: 351, Nvcsw: 281, Nivcsw: 3118, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/13/25 09:06:46.329 + > Enter [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/13/25 09:06:39.393 < Exit [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/13/25 09:06:42.454 (3.062s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/13/25 09:06:42.454 [FAILED] Warning: resource namespaces/head-5073ae15-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "ceph-pool-r2-csi-rbd" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x140001bc9e0>: exit status 1 { ProcessState: { pid: 11922, status: 256, rusage: { Utime: { Sec: 0, Usec: 297871, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 61792, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 68583424, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 4995, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 49, Msgrcv: 56, Nsignals: 351, Nvcsw: 281, Nivcsw: 3118, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/13/25 09:06:46.329 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/13/25 09:06:46.329 (3.875s) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/13/25 09:06:46.329 The list of pods is empty; nothing to dump. < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/13/25 09:06:52.338 (6.01s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:112 @ 10/13/25 09:06:52.339 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:147 @ 10/13/25 09:06:52.339 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:190 @ 10/13/25 09:06:52.339 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:196 @ 10/13/25 09:06:52.339 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:204 @ 10/13/25 09:06:52.339 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:224 @ 10/13/25 09:06:52.339 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:236 @ 10/13/25 09:06:52.339 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:258 @ 10/13/25 09:06:52.339 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:284 @ 10/13/25 09:06:52.339 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:294 @ 10/13/25 09:06:52.339 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:309 @ 10/13/25 09:06:52.34 + + + > Enter [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/13/25 09:06:52.34 < Exit [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/13/25 09:06:54.312 (1.972s) > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 09:06:54.312 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 09:06:54.312 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/13/25 09:06:54.312 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/13/25 09:07:00.494 (6.182s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 09:07:00.494 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 09:07:00.494 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1013 09:18:02.429841 11941 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nW1013 09:18:33.846318 11941 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: Get \"https://api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-restore-force/virtualmachines?fieldSelector=metadata.name%3Dhead-5073ae15-vm-restore-force&resourceVersion=58857\": dial tcp: lookup api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com: i/o timeout\nE1013 09:18:33.846497 11941 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: Get \\\"https://api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-restore-force/virtualmachines?fieldSelector=metadata.name%3Dhead-5073ae15-vm-restore-force&resourceVersion=58857\\\": dial tcp: lookup api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com: i/o timeout\"\nW1013 09:19:05.744484 11941 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: Get \"https://api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-restore-force/virtualmachines?fieldSelector=metadata.name%3Dhead-5073ae15-vm-restore-force&resourceVersion=58857\": dial tcp: lookup api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com: i/o timeout\nE1013 09:19:05.744760 11941 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: Get \\\"https://api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-restore-force/virtualmachines?fieldSelector=metadata.name%3Dhead-5073ae15-vm-restore-force&resourceVersion=58857\\\": dial tcp: lookup api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com: i/o timeout\"\nW1013 09:19:40.252514 11941 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: Get \"https://api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-restore-force/virtualmachines?fieldSelector=metadata.name%3Dhead-5073ae15-vm-restore-force&resourceVersion=58857\": dial tcp: lookup api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com: i/o timeout\nE1013 09:19:40.252663 11941 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: Get \\\"https://api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-resto... Gomega truncated this representation as it exceeds 'format.MaxLength'. Consider having the object provide a custom 'GomegaStringer' representation or adjust the parameters in Gomega's 'format' package. Learn more here: https://onsi.github.io/gomega/#adjusting-output to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 @ 10/13/25 09:23:42.682 + > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 09:07:00.494 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 09:07:00.494 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/13/25 09:07:00.494 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/13/25 09:07:00.494 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/13/25 09:23:42.682 (16m42.141s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1013 09:18:02.429841 11941 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nW1013 09:18:33.846318 11941 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: Get \"https://api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-restore-force/virtualmachines?fieldSelector=metadata.name%3Dhead-5073ae15-vm-restore-force&resourceVersion=58857\": dial tcp: lookup api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com: i/o timeout\nE1013 09:18:33.846497 11941 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: Get \\\"https://api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-restore-force/virtualmachines?fieldSelector=metadata.name%3Dhead-5073ae15-vm-restore-force&resourceVersion=58857\\\": dial tcp: lookup api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com: i/o timeout\"\nW1013 09:19:05.744484 11941 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: Get \"https://api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-restore-force/virtualmachines?fieldSelector=metadata.name%3Dhead-5073ae15-vm-restore-force&resourceVersion=58857\": dial tcp: lookup api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com: i/o timeout\nE1013 09:19:05.744760 11941 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: Get \\\"https://api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-restore-force/virtualmachines?fieldSelector=metadata.name%3Dhead-5073ae15-vm-restore-force&resourceVersion=58857\\\": dial tcp: lookup api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com: i/o timeout\"\nW1013 09:19:40.252514 11941 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: Get \"https://api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-restore-force/virtualmachines?fieldSelector=metadata.name%3Dhead-5073ae15-vm-restore-force&resourceVersion=58857\": dial tcp: lookup api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com: i/o timeout\nE1013 09:19:40.252663 11941 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: Get \\\"https://api.dvp-e2e-clean-sds-20251013-081321-32554.clean-sds-20251013-081321-32554.e2e.virtlab.flant.com/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-resto... Gomega truncated this representation as it exceeds 'format.MaxLength'. Consider having the object provide a custom 'GomegaStringer' representation or adjust the parameters in Gomega's 'format' package. Learn more here: https://onsi.github.io/gomega/#adjusting-output to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 @ 10/13/25 09:23:42.682 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/13/25 09:23:42.682 (16m42.142s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 09:23:42.682 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 09:23:48.55 (5.867s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:131 @ 10/13/25 09:23:48.55 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:177 @ 10/13/25 09:23:48.55 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:341 @ 10/13/25 09:23:48.55 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:367 @ 10/13/25 09:23:48.55 + + + + > Enter [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/13/25 09:23:48.55 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:73 @ 10/13/25 09:23:48.551 < Exit [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/13/25 09:23:48.551 (0s) > Enter [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/13/25 09:23:48.551 < Exit [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/13/25 09:23:48.551 (0s) > Enter [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/13/25 09:23:48.551 < Exit [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/13/25 09:23:48.551 (0s) + + + > Enter [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/13/25 09:23:48.551 < Exit [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/13/25 09:23:50.57 (2.018s) > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 09:23:50.57 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 09:23:50.57 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/13/25 09:23:50.57 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/13/25 09:23:57.715 (7.146s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 09:23:57.715 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 09:23:57.715 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 09:23:57.716 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 09:23:57.716 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/13/25 09:23:57.716 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:91 @ 10/13/25 09:23:57.716 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/13/25 09:24:01.852 (4.136s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 09:24:01.852 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 09:24:01.852 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=WaitForFirstConsumer'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-blank-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=WaitForFirstConsumer' --timeout=16m40s\nstderr: W1013 09:35:17.078932 13200 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: stream error: stream ID 15; INTERNAL_ERROR; received from peer\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-blank-not-existing-vmclass-with-changing\n\nwaited for: 'jsonpath={.status.phase}=WaitForFirstConsumer'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=WaitForFirstConsumer' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-not-existing-vmclass-with-changing\n\nwaited for: 'jsonpath={.status.phase}=WaitForFirstConsumer'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:103 @ 10/13/25 09:40:44.023 + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 09:24:01.852 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 09:24:01.852 (0s) > Enter [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/13/25 09:24:01.852 STEP: VDs should be in WaitForFirstConsumer phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:102 @ 10/13/25 09:24:01.852 [FAILED] should observe resources in ''jsonpath={.status.phase}=WaitForFirstConsumer'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-blank-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=WaitForFirstConsumer' --timeout=16m40s\nstderr: W1013 09:35:17.078932 13200 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: stream error: stream ID 15; INTERNAL_ERROR; received from peer\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-blank-not-existing-vmclass-with-changing\n\nwaited for: 'jsonpath={.status.phase}=WaitForFirstConsumer'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=WaitForFirstConsumer' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-not-existing-vmclass-with-changing\n\nwaited for: 'jsonpath={.status.phase}=WaitForFirstConsumer'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:103 @ 10/13/25 09:40:44.023 < Exit [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/13/25 09:40:44.024 (16m42.201s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 09:40:44.024 The list of pods is empty; nothing to dump. < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 09:40:50.472 (6.448s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/13/25 09:40:50.473 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:126 @ 10/13/25 09:40:50.473 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:140 @ 10/13/25 09:40:50.473 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:152 @ 10/13/25 09:40:50.473 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:157 @ 10/13/25 09:40:50.473 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:163 @ 10/13/25 09:40:50.473 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:176 @ 10/13/25 09:40:50.473 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:181 @ 10/13/25 09:40:50.473 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:187 @ 10/13/25 09:40:50.473 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:202 @ 10/13/25 09:40:50.473 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:216 @ 10/13/25 09:40:50.474 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:240 @ 10/13/25 09:40:50.474 + + + > Enter [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/13/25 09:40:50.474 < Exit [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/13/25 09:40:52.484 (2.01s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/13/25 09:40:52.484 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/13/25 09:41:01.885 (9.401s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 09:41:01.885 < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 09:41:01.885 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:6, cap:6>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-b-not-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-b-not-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-selector --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-selector\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-d --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-d\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-affinity --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-affinity\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-c-and-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-c-and-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-a-not-b --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-a-not-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:103 @ 10/13/25 09:57:53.977 + > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/13/25 09:41:01.885 STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/13/25 09:41:01.885 END STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/13/25 09:41:08.068 (6.183s) STEP: `VirtualMachineClasses` should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:95 @ 10/13/25 09:41:08.068 END STEP: `VirtualMachineClasses` should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:95 @ 10/13/25 09:41:11.34 (3.271s) STEP: `VirtualDisks` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:102 @ 10/13/25 09:41:11.34 END STEP: `VirtualDisks` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:102 @ 10/13/25 09:57:53.977 (16m42.598s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:6, cap:6>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-b-not-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-b-not-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-selector --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-selector\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-d --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-d\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-affinity --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-affinity\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-c-and-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-c-and-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-a-not-b --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-a-not-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:103 @ 10/13/25 09:57:53.977 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/13/25 09:57:53.978 (16m52.054s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 09:57:53.978 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 09:58:00.408 (6.43s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:120 @ 10/13/25 09:58:00.408 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:283 @ 10/13/25 09:58:00.409 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:372 @ 10/13/25 09:58:00.409 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:467 @ 10/13/25 09:58:00.409 + + + > Enter [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/13/25 09:58:00.409 < Exit [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/13/25 09:58:00.41 (1ms) > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:58:00.41 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:58:00.41 (0s) > Enter [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/13/25 09:58:00.41 < Exit [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/13/25 09:58:02.139 (1.729s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:58:02.139 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:58:02.139 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:58:02.14 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:58:02.14 (0s) > Enter [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/13/25 09:58:02.14 STEP: Project should be deployed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:73 @ 10/13/25 09:58:02.14 < Exit [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/13/25 09:58:04.932 (2.793s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:58:04.932 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:58:04.932 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:58:04.933 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:58:04.933 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/13/25 09:58:04.933 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/13/25 09:58:08.113 (3.181s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:58:08.114 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:58:08.114 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:58:08.114 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:58:08.114 (0s) > Enter [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/13/25 09:58:08.114 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/13/25 09:58:08.114 < Exit [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/13/25 09:58:15.049 (6.935s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:58:15.049 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 09:58:15.049 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-cvi --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-cvi\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/13/25 10:14:57.142 + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:58:15.05 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 09:58:15.05 (0s) > Enter [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/13/25 09:58:15.05 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/13/25 09:58:15.05 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-cvi --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-cvi\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/13/25 10:14:57.142 < Exit [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/13/25 10:14:57.142 (16m42.087s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 10:14:57.142 The list of pods is empty; nothing to dump. < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 10:15:03.609 (6.467s) > Enter [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/13/25 10:15:03.609 STEP: Delete manifests - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:50 @ 10/13/25 10:15:03.61 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 10:15:03.61 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 10:15:06.711 (3.101s) < Exit [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/13/25 10:15:06.711 (3.101s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/13/25 10:15:06.712 + + + + > Enter [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/13/25 10:15:06.712 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:51 @ 10/13/25 10:15:06.712 < Exit [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/13/25 10:15:06.712 (0s) > Enter [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/13/25 10:15:06.712 < Exit [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/13/25 10:15:06.712 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:82 @ 10/13/25 10:15:06.712 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:94 @ 10/13/25 10:15:06.712 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:106 @ 10/13/25 10:15:06.712 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:116 @ 10/13/25 10:15:06.712 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:132 @ 10/13/25 10:15:06.713 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:151 @ 10/13/25 10:15:06.713 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:157 @ 10/13/25 10:15:06.713 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:189 @ 10/13/25 10:15:06.713 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:216 @ 10/13/25 10:15:06.713 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:252 @ 10/13/25 10:15:06.713 + + + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/13/25 10:15:06.713 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/13/25 10:15:08.746 (2.032s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/13/25 10:15:08.746 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/13/25 10:15:12.679 (3.933s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 10:15:12.679 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 10:15:12.679 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/13/25 10:15:12.679 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:97 @ 10/13/25 10:15:12.679 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/13/25 10:15:26.491 (13.812s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 10:15:26.491 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 10:15:26.491 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-a --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-b --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 @ 10/13/25 10:32:09.086 + > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/13/25 10:15:26.491 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 @ 10/13/25 10:15:26.491 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-a --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-b --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 @ 10/13/25 10:32:09.086 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/13/25 10:32:09.086 (16m42.589s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 10:32:09.086 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 10:32:15.453 (6.367s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:118 @ 10/13/25 10:32:15.453 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:129 @ 10/13/25 10:32:15.453 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:142 @ 10/13/25 10:32:15.453 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:166 @ 10/13/25 10:32:15.453 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:174 @ 10/13/25 10:32:15.453 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:179 @ 10/13/25 10:32:15.454 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:187 @ 10/13/25 10:32:15.454 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:198 @ 10/13/25 10:32:15.454 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:209 @ 10/13/25 10:32:15.454 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:222 @ 10/13/25 10:32:15.454 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:228 @ 10/13/25 10:32:15.454 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:240 @ 10/13/25 10:32:15.454 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:250 @ 10/13/25 10:32:15.454 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:258 @ 10/13/25 10:32:15.454 + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1013 10:37:00.604437 17978 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: stream error: stream ID 13; INTERNAL_ERROR; received from peer\") has prevented the request from succeeding\nW1013 10:37:02.785716 17978 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding\nE1013 10:37:02.785869 17978 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: an error on the server (\\\"<html>\\\\r\\\\n<head><title>502 Bad Gateway</title></head>\\\\r\\\\n<body>\\\\r\\\\n<center><h1>502 Bad Gateway</h1></center>\\\\r\\\\n<hr><center>nginx</center>\\\\r\\\\n</body>\\\\r\\\\n</html>\\\") has prevented the request from succeeding\"\nW1013 10:37:05.591222 17978 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding\nE1013 10:37:05.591270 17978 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: an error on the server (\\\"<html>\\\\r\\\\n<head><title>502 Bad Gateway</title></head>\\\\r\\\\n<body>\\\\r\\\\n<center><h1>502 Bad Gateway</h1></center>\\\\r\\\\n<hr><center>nginx</center>\\\\r\\\\n</body>\\\\r\\\\n</html>\\\") has prevented the request from succeeding\"\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1013 10:37:00.600947 17977 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: stream error: stream ID 13; INTERNAL_ERROR; received from peer\") has prevented the request from succeeding\nW1013 10:37:02.988597 17977 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding\nE1013 10:37:02.988743 17977 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: an error on the server (\\\"<html>\\\\r\\\\n<head><title>502 Bad Gateway</title></head>\\\\r\\\\n<body>\\\\r\\\\n<center><h1>502 Bad Gateway</h1></center>\\\\r\\\\n<hr><center>nginx</center>\\\\r\\\\n</body>\\\\r\\\\n</html>\\\") has prevented the request from succeeding\"\nW1013 10:37:05.751151 17977 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to l... Gomega truncated this representation as it exceeds 'format.MaxLength'. Consider having the object provide a custom 'GomegaStringer' representation or adjust the parameters in Gomega's 'format' package. Learn more here: https://onsi.github.io/gomega/#adjusting-output to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 @ 10/13/25 10:49:04.407 + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/13/25 10:32:15.454 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/13/25 10:32:17.483 (2.029s) > Enter [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/13/25 10:32:17.483 < Exit [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/13/25 10:32:22.381 (4.898s) > Enter [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/13/25 10:32:22.381 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 @ 10/13/25 10:32:22.382 [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1013 10:37:00.604437 17978 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: stream error: stream ID 13; INTERNAL_ERROR; received from peer\") has prevented the request from succeeding\nW1013 10:37:02.785716 17978 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding\nE1013 10:37:02.785869 17978 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: an error on the server (\\\"<html>\\\\r\\\\n<head><title>502 Bad Gateway</title></head>\\\\r\\\\n<body>\\\\r\\\\n<center><h1>502 Bad Gateway</h1></center>\\\\r\\\\n<hr><center>nginx</center>\\\\r\\\\n</body>\\\\r\\\\n</html>\\\") has prevented the request from succeeding\"\nW1013 10:37:05.591222 17978 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding\nE1013 10:37:05.591270 17978 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: an error on the server (\\\"<html>\\\\r\\\\n<head><title>502 Bad Gateway</title></head>\\\\r\\\\n<body>\\\\r\\\\n<center><h1>502 Bad Gateway</h1></center>\\\\r\\\\n<hr><center>nginx</center>\\\\r\\\\n</body>\\\\r\\\\n</html>\\\") has prevented the request from succeeding\"\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1013 10:37:00.600947 17977 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: stream error: stream ID 13; INTERNAL_ERROR; received from peer\") has prevented the request from succeeding\nW1013 10:37:02.988597 17977 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: an error on the server (\"<html>\\r\\n<head><title>502 Bad Gateway</title></head>\\r\\n<body>\\r\\n<center><h1>502 Bad Gateway</h1></center>\\r\\n<hr><center>nginx</center>\\r\\n</body>\\r\\n</html>\") has prevented the request from succeeding\nE1013 10:37:02.988743 17977 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: an error on the server (\\\"<html>\\\\r\\\\n<head><title>502 Bad Gateway</title></head>\\\\r\\\\n<body>\\\\r\\\\n<center><h1>502 Bad Gateway</h1></center>\\\\r\\\\n<hr><center>nginx</center>\\\\r\\\\n</body>\\\\r\\\\n</html>\\\") has prevented the request from succeeding\"\nW1013 10:37:05.751151 17977 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to l... Gomega truncated this representation as it exceeds 'format.MaxLength'. Consider having the object provide a custom 'GomegaStringer' representation or adjust the parameters in Gomega's 'format' package. Learn more here: https://onsi.github.io/gomega/#adjusting-output to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 @ 10/13/25 10:49:04.407 < Exit [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/13/25 10:49:04.407 (16m42.077s) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/13/25 10:49:04.407 The list of pods is empty; nothing to dump. STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 10:49:11.257 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 10:49:15.695 (4.438s) < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/13/25 10:49:15.696 (11.288s) + + + > Enter [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/13/25 10:49:15.696 < Exit [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/13/25 10:49:17.722 (2.026s) > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 10:49:17.722 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 10:49:17.723 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/13/25 10:49:17.723 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/13/25 10:49:20.057 (2.334s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 10:49:20.057 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 10:49:20.057 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root --namespace head-5073ae15-end-to-end-vm-versions --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:69 @ 10/13/25 11:06:02.264 + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 10:49:20.057 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 10:49:20.057 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/13/25 10:49:20.057 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:68 @ 10/13/25 10:49:20.058 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root --namespace head-5073ae15-end-to-end-vm-versions --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:69 @ 10/13/25 11:06:02.264 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/13/25 11:06:02.264 (16m42.167s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 11:06:02.264 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 11:06:11.31 (9.046s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/13/25 11:06:11.311 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/13/25 11:06:11.311 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/13/25 11:06:11.311 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/13/25 11:06:11.311 + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/13/25 11:06:13.285 + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/13/25 11:06:11.311 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/13/25 11:06:13.285 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/13/25 11:06:13.285 (1.974s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/13/25 11:06:13.285 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/13/25 11:06:19.584 (6.299s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:100 @ 10/13/25 11:06:19.584 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:111 @ 10/13/25 11:06:19.584 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:122 @ 10/13/25 11:06:19.584 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:133 @ 10/13/25 11:06:19.584 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:144 @ 10/13/25 11:06:19.584 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:164 @ 10/13/25 11:06:19.585 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:185 @ 10/13/25 11:06:19.585 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:229 @ 10/13/25 11:06:19.585 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:290 @ 10/13/25 11:06:19.585 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:341 @ 10/13/25 11:06:19.585 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:367 @ 10/13/25 11:06:19.585 + + + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/13/25 11:06:19.585 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/13/25 11:06:21.587 (2.001s) > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 11:06:21.587 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 11:06:21.587 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/13/25 11:06:21.587 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/13/25 11:06:25.316 (3.729s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 11:06:25.316 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 11:06:25.316 (0s) + + + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 11:06:25.317 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 11:06:25.317 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/13/25 11:06:25.317 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:88 @ 10/13/25 11:06:25.317 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/13/25 11:06:30.579 (5.262s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 11:06:30.579 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 11:06:30.579 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-with-hotplug-standalone --namespace head-5073ae15-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/13/25 11:23:12.727 + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 11:06:30.579 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 11:06:30.579 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/13/25 11:06:30.579 STEP: VDs with consumers should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:99 @ 10/13/25 11:06:30.579 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-with-hotplug-standalone --namespace head-5073ae15-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/13/25 11:23:12.727 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/13/25 11:23:12.727 (16m42.12s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 11:23:12.727 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 11:23:19.046 (6.319s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/13/25 11:23:19.046 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/13/25 11:23:19.046 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/13/25 11:23:19.046 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/13/25 11:23:19.046 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/13/25 11:23:19.047 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/13/25 11:23:19.047 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/13/25 11:23:19.047 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/13/25 11:23:19.047 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/13/25 11:23:19.047 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/13/25 11:23:19.047 + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.047 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 11:23:19.047 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.047 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.048 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 11:23:19.048 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.048 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.048 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 11:23:19.048 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.048 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.048 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 11:23:19.049 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.049 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.049 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 11:23:19.049 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.049 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.049 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 11:23:19.049 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.049 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.049 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 11:23:19.049 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.049 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.049 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 11:23:19.049 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.05 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.05 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 11:23:19.05 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 11:23:19.05 (0s) + + + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/13/25 11:23:19.05 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/13/25 11:23:21.039 (1.989s) > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 11:23:21.039 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 11:23:21.039 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/13/25 11:23:21.039 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/13/25 11:23:23.851 (2.812s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 11:23:23.851 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 11:23:23.851 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 11:23:23.851 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 11:23:23.851 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/13/25 11:23:23.851 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:75 @ 10/13/25 11:23:23.851 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/13/25 11:23:28.59 (4.739s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 11:23:28.59 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 11:23:28.591 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-label-annotation --namespace head-5073ae15-end-to-end-vm-label-annotation --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-label-annotation\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 @ 10/13/25 11:40:10.706 + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 11:23:28.591 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 11:23:28.591 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/13/25 11:23:28.591 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 @ 10/13/25 11:23:28.591 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-label-annotation --namespace head-5073ae15-end-to-end-vm-label-annotation --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-label-annotation\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 @ 10/13/25 11:40:10.706 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/13/25 11:40:10.707 (16m42.108s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 11:40:10.707 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 11:40:17.466 (6.76s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/13/25 11:40:17.467 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/13/25 11:40:17.467 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/13/25 11:40:17.467 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/13/25 11:40:17.467 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/13/25 11:40:17.467 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/13/25 11:40:17.467 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/13/25 11:40:17.467 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/13/25 11:40:17.467 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/13/25 11:40:17.468 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/13/25 11:40:17.468 + + + + > Enter [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/13/25 11:40:17.468 [SKIPPED] Module SDN is disabled. Skipping all tests for module SDN. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:50 @ 10/13/25 11:40:18.134 < Exit [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/13/25 11:40:18.134 (666ms) > Enter [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/13/25 11:40:18.134 < Exit [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/13/25 11:40:18.134 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/13/25 11:40:18.134 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/13/25 11:40:18.134 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:111 @ 10/13/25 11:40:18.134 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:125 @ 10/13/25 11:40:18.134 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:141 @ 10/13/25 11:40:18.134 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:156 @ 10/13/25 11:40:18.135 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:169 @ 10/13/25 11:40:18.135 + + + > Enter [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/13/25 11:40:18.135 < Exit [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/13/25 11:40:20.148 (2.013s) > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/13/25 11:40:20.148 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/13/25 11:40:20.148 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/13/25 11:40:20.148 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/13/25 11:40:24.398 (4.25s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 11:40:24.398 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 11:40:24.399 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-safe --namespace head-5073ae15-end-to-end-vm-restore-safe --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-safe\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:107 @ 10/13/25 11:57:06.955 + > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/13/25 11:40:24.399 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/13/25 11:40:24.399 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/13/25 11:40:24.399 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/13/25 11:40:24.399 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/13/25 11:57:06.955 (16m42.548s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-safe --namespace head-5073ae15-end-to-end-vm-restore-safe --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-safe\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:107 @ 10/13/25 11:57:06.955 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/13/25 11:57:06.956 (16m42.548s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 11:57:06.956 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 11:57:13.478 (6.523s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:115 @ 10/13/25 11:57:13.478 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:158 @ 10/13/25 11:57:13.479 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:348 @ 10/13/25 11:57:13.479 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:374 @ 10/13/25 11:57:13.479 + + + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/13/25 11:57:13.479 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/13/25 11:57:15.496 (2.017s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:62 @ 10/13/25 11:57:15.496 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:62 @ 10/13/25 11:57:44.378 (28.882s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 11:57:44.378 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 11:57:44.378 (0s) + + + > Enter [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/13/25 11:57:44.378 < Exit [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/13/25 11:57:45.558 (1.18s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 11:57:45.558 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 11:57:45.559 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/13/25 11:57:45.559 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:102 @ 10/13/25 11:57:45.559 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/13/25 11:57:51.825 (6.267s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 11:57:51.825 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 11:57:51.826 (0s) + + + > Enter [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/13/25 11:57:51.826 STEP: CVIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:113 @ 10/13/25 11:57:51.826 < Exit [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/13/25 11:57:54.791 (2.965s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 11:57:54.791 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 11:57:54.791 (0s) + + + > Enter [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/13/25 11:57:54.791 STEP: VMClasses should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:124 @ 10/13/25 11:57:54.791 < Exit [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/13/25 11:57:57.524 (2.732s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 11:57:57.524 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 11:57:57.524 (0s) + + + [FAILED] Timed out after 347.531s. Expected success, but got an error: <*errors.errorString | 0x14000bfc0a0>: cannot patch VMIP "head-5073ae15-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip { s: "cannot patch VMIP \"head-5073ae15-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:138 @ 10/13/25 12:03:45.059 + > Enter [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/13/25 11:57:57.524 [FAILED] Timed out after 347.531s. Expected success, but got an error: <*errors.errorString | 0x14000bfc0a0>: cannot patch VMIP "head-5073ae15-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip { s: "cannot patch VMIP \"head-5073ae15-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:138 @ 10/13/25 12:03:45.059 < Exit [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/13/25 12:03:45.059 (5m47.533s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 12:03:45.06 The list of pods is empty; nothing to dump. < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 12:03:51.848 (6.788s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:141 @ 10/13/25 12:03:51.848 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:152 @ 10/13/25 12:03:51.848 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:162 @ 10/13/25 12:03:51.848 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:173 @ 10/13/25 12:03:51.848 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:184 @ 10/13/25 12:03:51.848 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:196 @ 10/13/25 12:03:51.848 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:220 @ 10/13/25 12:03:51.848 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:245 @ 10/13/25 12:03:51.848 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:268 @ 10/13/25 12:03:51.849 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:280 @ 10/13/25 12:03:51.849 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:299 @ 10/13/25 12:03:51.849 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:316 @ 10/13/25 12:03:51.849 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:341 @ 10/13/25 12:03:51.849 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:354 @ 10/13/25 12:03:51.849 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:358 @ 10/13/25 12:03:51.849 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:375 @ 10/13/25 12:03:51.849 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:388 @ 10/13/25 12:03:51.849 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:405 @ 10/13/25 12:03:51.849 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:444 @ 10/13/25 12:03:51.849 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:479 @ 10/13/25 12:03:51.849 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:497 @ 10/13/25 12:03:51.849 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:512 @ 10/13/25 12:03:51.85 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:528 @ 10/13/25 12:03:51.85 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:545 @ 10/13/25 12:03:51.85 + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x140001a3bc0>: the container "virtualization-controller" was restarted: virtualization-controller-7b6676cf8f-rcvh9 the container "virtualization-controller" was not found: virtualization-controller-7d77d9f5b5-8r9qb { errs: [ <*errors.joinError | 0x140001a3b90>{ errs: [ <*errors.errorString | 0x14000d0f9e0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-7b6676cf8f-rcvh9", }, ], }, <*errors.errorString | 0x14000d0fa10>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-7d77d9f5b5-8r9qb", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/13/25 12:03:52.238 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 12:03:51.85 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 12:03:51.85 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 12:03:51.85 [FAILED] Expected success, but got an error: <*errors.joinError | 0x140001a3bc0>: the container "virtualization-controller" was restarted: virtualization-controller-7b6676cf8f-rcvh9 the container "virtualization-controller" was not found: virtualization-controller-7d77d9f5b5-8r9qb { errs: [ <*errors.joinError | 0x140001a3b90>{ errs: [ <*errors.errorString | 0x14000d0f9e0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-7b6676cf8f-rcvh9", }, ], }, <*errors.errorString | 0x14000d0fa10>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-7d77d9f5b5-8r9qb", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/13/25 12:03:52.238 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 12:03:52.238 (388ms) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/13/25 12:03:52.238 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/13/25 12:04:03.96 (11.721s) + + + \ No newline at end of file diff --git a/artifacts/clean-sds-20251013-161219-2917/junit.xml b/artifacts/clean-sds-20251013-161219-2917/junit.xml new file mode 100644 index 0000000000..ca88f5b246 --- /dev/null +++ b/artifacts/clean-sds-20251013-161219-2917/junit.xml @@ -0,0 +1,883 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 16:34:26.877 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 16:34:31.908 (5.031s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 16:34:31.908 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 16:34:31.908 (0s) + + + > Enter [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/13/25 16:34:31.911 < Exit [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/13/25 16:34:33.469 (1.558s) > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 16:34:33.469 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 16:34:33.469 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/13/25 16:34:33.469 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/13/25 16:34:36.337 (2.868s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 16:34:36.337 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 16:34:36.337 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root --namespace head-5073ae15-end-to-end-vm-versions --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1013 16:38:33.502948 38228 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: stream error: stream ID 13; INTERNAL_ERROR; received from peer\") has prevented the request from succeeding\nW1013 16:38:35.459641 38228 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 16:38:35.460195 38228 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 16:38:38.763947 38228 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 16:38:38.764045 38228 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 16:38:42.560065 38228 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 16:38:42.560165 38228 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 16:38:49.238681 38228 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 16:38:49.238764 38228 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 16:39:02.497829 38228 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 16:39:02.497871 38228 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 16:39:29.180952 38228 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 16:39:29.181003 38228 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 16:40:22.594179 38228 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 16:40:22.594242 38228 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 16:41:09.492747 38228 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 16:41:09.492849 38228 reflector.go:166]... Gomega truncated this representation as it exceeds 'format.MaxLength'. Consider having the object provide a custom 'GomegaStringer' representation or adjust the parameters in Gomega's 'format' package. Learn more here: https://onsi.github.io/gomega/#adjusting-output to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:69 @ 10/13/25 16:51:18.472 + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 16:34:36.338 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 16:34:36.338 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/13/25 16:34:36.338 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:68 @ 10/13/25 16:34:36.338 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root --namespace head-5073ae15-end-to-end-vm-versions --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1013 16:38:33.502948 38228 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: stream error: stream ID 13; INTERNAL_ERROR; received from peer\") has prevented the request from succeeding\nW1013 16:38:35.459641 38228 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 16:38:35.460195 38228 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 16:38:38.763947 38228 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 16:38:38.764045 38228 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 16:38:42.560065 38228 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 16:38:42.560165 38228 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 16:38:49.238681 38228 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 16:38:49.238764 38228 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 16:39:02.497829 38228 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 16:39:02.497871 38228 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 16:39:29.180952 38228 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 16:39:29.181003 38228 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 16:40:22.594179 38228 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 16:40:22.594242 38228 reflector.go:166] \"Unhandled Error\" err=\"k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *unstructured.Unstructured: failed to list *unstructured.Unstructured: the server could not find the requested resource\"\nW1013 16:41:09.492747 38228 reflector.go:569] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *unstructured.Unstructured: the server could not find the requested resource\nE1013 16:41:09.492849 38228 reflector.go:166]... Gomega truncated this representation as it exceeds 'format.MaxLength'. Consider having the object provide a custom 'GomegaStringer' representation or adjust the parameters in Gomega's 'format' package. Learn more here: https://onsi.github.io/gomega/#adjusting-output to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:69 @ 10/13/25 16:51:18.472 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/13/25 16:51:18.473 (16m42.101s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 16:51:18.473 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-versions -o yaml -l 'testcase=vm-versions' --show-managed-fields=true error: exit status 1 stderr: E1013 16:51:19.309624 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:19.559422 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:19.814561 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:20.064211 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:20.314027 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:20.556970 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:20.803408 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:21.060371 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:21.308065 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:21.554319 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:21.806858 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:22.085360 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:22.340084 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:22.583804 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:22.832663 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:23.081768 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:23.338333 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:23.586835 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:23.829027 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:24.071898 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:24.318551 39129 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-versions -o json -l 'testcase=vm-versions' stderr: E1013 16:51:25.044089 39134 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:25.289130 39134 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:25.543421 39134 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:25.789535 39134 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:26.046749 39134 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:26.294206 39134 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=vm-versions", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 16:51:26.427 (7.954s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/13/25 16:51:26.427 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/13/25 16:51:26.427 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/13/25 16:51:26.427 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/13/25 16:51:26.427 + + + + > Enter [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/13/25 16:51:26.428 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:41 @ 10/13/25 16:51:26.428 < Exit [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/13/25 16:51:26.428 (0s) > Enter [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/13/25 16:51:26.428 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 16:51:26.428 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 16:51:33.221 (6.793s) [FAILED] cannot delete test case resources cmd: kubectl delete virtualmachineoperations.virtualization.deckhouse.io -l 'testcase=vm-migration-cancel' stderr: E1013 16:51:30.773548 39136 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:31.023230 39136 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:31.265703 39136 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:31.517625 39136 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:31.758902 39136 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:31.997355 39136 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:32.240810 39136 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:32.502542 39136 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:32.739305 39136 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:32.977258 39136 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:33.217332 39136 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Unexpected error: <*exec.ExitError | 0x1400061cf60>: exit status 1 { ProcessState: { pid: 39136, status: 256, rusage: { Utime: { Sec: 0, Usec: 84810, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 66644, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 49512448, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3832, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 33, Msgrcv: 32, Nsignals: 111, Nvcsw: 82, Nivcsw: 2409, }, }, Stderr: nil, } occurred In [AfterEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:626 @ 10/13/25 16:51:33.221 < Exit [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/13/25 16:51:33.222 (6.794s) + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-sizing-policy Unexpected error: <*exec.ExitError | 0x1400061d680>: exit status 1 { ProcessState: { pid: 39138, status: 256, rusage: { Utime: { Sec: 0, Usec: 34096, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 17343, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47529984, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3713, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 13, Nsignals: 52, Nvcsw: 35, Nivcsw: 694, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:64 @ 10/13/25 16:51:34.093 + > Enter [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/13/25 16:51:33.222 [FAILED] kubectl create namespace head-5073ae15-end-to-end-sizing-policy Unexpected error: <*exec.ExitError | 0x1400061d680>: exit status 1 { ProcessState: { pid: 39138, status: 256, rusage: { Utime: { Sec: 0, Usec: 34096, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 17343, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47529984, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3713, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 13, Nsignals: 52, Nvcsw: 35, Nivcsw: 694, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:64 @ 10/13/25 16:51:34.093 < Exit [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/13/25 16:51:34.093 (871ms) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 16:51:34.093 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-sizing-policy -o yaml -l 'testcase=sizing-policy' --show-managed-fields=true error: exit status 1 stderr: E1013 16:51:34.824420 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:35.063842 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:35.311677 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:35.550468 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:35.792958 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:36.031667 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:36.274170 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:36.552454 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:36.801041 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:37.040319 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:37.283295 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:37.524915 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:37.767558 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:38.007801 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:38.249496 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:38.498379 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:38.747951 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:38.987251 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:39.225466 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:39.463743 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:39.701747 39139 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-sizing-policy -o json -l 'testcase=sizing-policy' stderr: E1013 16:51:40.813984 39142 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:41.065403 39142 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:41.316777 39142 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:41.561209 39142 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:41.839286 39142 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:42.083641 39142 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=sizing-policy", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 16:51:42.211 (8.118s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/13/25 16:51:42.211 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/13/25 16:51:42.211 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/13/25 16:51:42.211 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:126 @ 10/13/25 16:51:42.211 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:140 @ 10/13/25 16:51:42.211 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:152 @ 10/13/25 16:51:42.211 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:157 @ 10/13/25 16:51:42.211 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:163 @ 10/13/25 16:51:42.212 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:176 @ 10/13/25 16:51:42.212 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:181 @ 10/13/25 16:51:42.212 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:187 @ 10/13/25 16:51:42.212 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:202 @ 10/13/25 16:51:42.212 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:216 @ 10/13/25 16:51:42.212 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:240 @ 10/13/25 16:51:42.212 + + + [FAILED] error: error validating "/tmp/testdata/importer-network-policy/project": error validating data: failed to download openapi: the server could not find the requested resource; if you choose to ignore these errors, turn validation off with --validate=false Expected <bool>: false to equal <bool>: true In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:69 @ 10/13/25 16:51:43.298 + > Enter [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/13/25 16:51:42.212 < Exit [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/13/25 16:51:42.213 (1ms) > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 16:51:42.213 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 16:51:42.213 (0s) > Enter [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/13/25 16:51:42.213 [FAILED] error: error validating "/tmp/testdata/importer-network-policy/project": error validating data: failed to download openapi: the server could not find the requested resource; if you choose to ignore these errors, turn validation off with --validate=false Expected <bool>: false to equal <bool>: true In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:69 @ 10/13/25 16:51:43.298 < Exit [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/13/25 16:51:43.298 (1.086s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 16:51:43.298 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-importer-network-policy -o yaml -l 'testcase=importer-network-policy' --show-managed-fields=true error: exit status 1 stderr: E1013 16:51:44.208123 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:44.449440 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:44.694164 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:44.938742 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:45.209775 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:45.457188 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:45.704230 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:45.950172 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:46.196484 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:46.440432 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:46.689464 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:46.932961 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:47.288278 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:47.535518 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:47.787099 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:48.037759 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:48.285938 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:48.532591 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:48.781877 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:49.026836 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:49.301649 39149 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-importer-network-policy -o json -l 'testcase=importer-network-policy' stderr: E1013 16:51:50.222482 39150 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:50.466668 39150 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:50.709765 39150 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:50.984256 39150 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:51.235467 39150 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:51.474435 39150 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=importer-network-policy", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 16:51:51.606 (8.308s) > Enter [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/13/25 16:51:51.606 STEP: Delete manifests - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:50 @ 10/13/25 16:51:51.606 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 16:51:51.606 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 16:51:53.956 (2.35s) < Exit [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/13/25 16:51:53.956 (2.35s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/13/25 16:51:53.956 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/13/25 16:51:53.956 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/13/25 16:51:53.956 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/13/25 16:51:53.956 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/13/25 16:51:53.957 + + + [FAILED] Unexpected error: <*exec.ExitError | 0x140005cafc0>: exit status 1 { ProcessState: { pid: 39158, status: 256, rusage: { Utime: { Sec: 0, Usec: 91259, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 100640, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 50266112, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3895, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 33, Msgrcv: 33, Nsignals: 117, Nvcsw: 106, Nivcsw: 2637, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:65 @ 10/13/25 16:51:57.196 + > Enter [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/13/25 16:51:53.957 [FAILED] Unexpected error: <*exec.ExitError | 0x140005cafc0>: exit status 1 { ProcessState: { pid: 39158, status: 256, rusage: { Utime: { Sec: 0, Usec: 91259, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 100640, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 50266112, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3895, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 33, Msgrcv: 33, Nsignals: 117, Nvcsw: 106, Nivcsw: 2637, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:65 @ 10/13/25 16:51:57.196 < Exit [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/13/25 16:51:57.196 (3.239s) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/13/25 16:51:57.196 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-image-hotplug -o yaml -l 'testcase=image-hotplug' --show-managed-fields=true error: exit status 1 stderr: E1013 16:51:57.888709 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:58.137228 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:58.382649 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:58.630589 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:58.935348 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:59.178944 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:59.437653 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:59.676739 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:51:59.920025 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:00.164450 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:00.410376 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:00.651692 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:00.895438 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:01.139866 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:01.387317 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:01.635605 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:01.880252 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:02.121423 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:02.372513 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:02.613510 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:02.854837 39160 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-image-hotplug -o json -l 'testcase=image-hotplug' stderr: E1013 16:52:03.815814 39162 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:04.088600 39162 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:04.340425 39162 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:04.586454 39162 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:04.839961 39162 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:05.086153 39162 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=image-hotplug", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/13/25 16:52:05.215 (8.019s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:112 @ 10/13/25 16:52:05.215 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:147 @ 10/13/25 16:52:05.215 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:190 @ 10/13/25 16:52:05.215 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:196 @ 10/13/25 16:52:05.215 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:204 @ 10/13/25 16:52:05.215 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:224 @ 10/13/25 16:52:05.215 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:236 @ 10/13/25 16:52:05.215 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:258 @ 10/13/25 16:52:05.215 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:284 @ 10/13/25 16:52:05.216 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:294 @ 10/13/25 16:52:05.216 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:309 @ 10/13/25 16:52:05.216 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-affinity-toleration Unexpected error: <*exec.ExitError | 0x1400061c460>: exit status 1 { ProcessState: { pid: 39163, status: 256, rusage: { Utime: { Sec: 0, Usec: 29916, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 11071, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47579136, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3716, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 54, Nvcsw: 24, Nivcsw: 619, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:69 @ 10/13/25 16:52:05.841 + > Enter [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/13/25 16:52:05.216 [FAILED] kubectl create namespace head-5073ae15-end-to-end-affinity-toleration Unexpected error: <*exec.ExitError | 0x1400061c460>: exit status 1 { ProcessState: { pid: 39163, status: 256, rusage: { Utime: { Sec: 0, Usec: 29916, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 11071, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47579136, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3716, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 54, Nvcsw: 24, Nivcsw: 619, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:69 @ 10/13/25 16:52:05.841 < Exit [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/13/25 16:52:05.842 (625ms) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 16:52:05.842 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-affinity-toleration -o yaml -l 'testcase=affinity-toleration' --show-managed-fields=true error: exit status 1 stderr: E1013 16:52:06.874090 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:07.119185 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:07.374925 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:07.620563 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:07.873557 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:08.122515 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:08.378308 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:08.622811 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:08.873057 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:09.120339 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:09.375921 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:09.623559 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:09.886697 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:10.136830 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:10.387514 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:10.633502 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:10.882792 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:11.127954 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:11.374199 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:11.655435 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:11.902518 39164 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-affinity-toleration -o json -l 'testcase=affinity-toleration' stderr: E1013 16:52:12.889110 39165 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:13.134571 39165 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:13.383279 39165 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:13.633152 39165 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:13.882171 39165 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:14.189950 39165 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=affinity-toleration", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 16:52:14.321 (8.479s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/13/25 16:52:14.321 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:120 @ 10/13/25 16:52:14.321 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:283 @ 10/13/25 16:52:14.322 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:372 @ 10/13/25 16:52:14.322 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:467 @ 10/13/25 16:52:14.322 + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.322 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 16:52:14.322 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.322 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.322 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 16:52:14.323 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.323 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.323 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 16:52:14.323 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.323 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.323 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 16:52:14.323 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.323 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.323 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 16:52:14.323 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.323 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.323 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 16:52:14.323 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.324 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.324 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 16:52:14.324 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.324 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.324 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 16:52:14.324 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.324 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.324 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 16:52:14.325 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 16:52:14.325 (0s) + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-migration Unexpected error: <*exec.ExitError | 0x1400065a040>: exit status 1 { ProcessState: { pid: 39167, status: 256, rusage: { Utime: { Sec: 0, Usec: 37561, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 19224, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46874624, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3672, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 13, Nsignals: 50, Nvcsw: 25, Nivcsw: 649, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:42 @ 10/13/25 16:52:14.851 + > Enter [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/13/25 16:52:14.325 [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-migration Unexpected error: <*exec.ExitError | 0x1400065a040>: exit status 1 { ProcessState: { pid: 39167, status: 256, rusage: { Utime: { Sec: 0, Usec: 37561, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 19224, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46874624, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3672, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 13, Nsignals: 50, Nvcsw: 25, Nivcsw: 649, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:42 @ 10/13/25 16:52:14.851 < Exit [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/13/25 16:52:14.851 (526ms) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 16:52:14.851 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-migration -o yaml -l 'testcase=vm-migration' --show-managed-fields=true error: exit status 1 stderr: E1013 16:52:15.922981 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:16.168795 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:16.417130 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:16.662787 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:16.944507 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:17.189774 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:17.442864 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:17.689077 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:17.941072 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:18.193839 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:18.444339 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:18.690304 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:18.941662 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:19.188097 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:19.437037 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:19.683844 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:19.934964 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:20.180482 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:20.428324 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:20.675492 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:20.954324 39168 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-migration -o json -l 'testcase=vm-migration' stderr: E1013 16:52:21.908902 39169 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:22.179679 39169 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:22.423847 39169 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:22.667160 39169 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:22.912752 39169 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:23.154102 39169 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=vm-migration", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 16:52:23.284 (8.432s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/13/25 16:52:23.284 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:86 @ 10/13/25 16:52:23.284 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:100 @ 10/13/25 16:52:23.284 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:116 @ 10/13/25 16:52:23.284 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:131 @ 10/13/25 16:52:23.284 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-disk-attachment Unexpected error: <*exec.ExitError | 0x14000a24f80>: exit status 1 { ProcessState: { pid: 39171, status: 256, rusage: { Utime: { Sec: 0, Usec: 40457, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 15449, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47661056, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3717, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 14, Nsignals: 55, Nvcsw: 51, Nivcsw: 617, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:61 @ 10/13/25 16:52:23.926 + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/13/25 16:52:23.284 [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-disk-attachment Unexpected error: <*exec.ExitError | 0x14000a24f80>: exit status 1 { ProcessState: { pid: 39171, status: 256, rusage: { Utime: { Sec: 0, Usec: 40457, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 15449, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47661056, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3717, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 14, Nsignals: 55, Nvcsw: 51, Nivcsw: 617, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:61 @ 10/13/25 16:52:23.926 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/13/25 16:52:23.926 (642ms) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 16:52:23.926 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-disk-attachment -o yaml -l 'testcase=vm-disk-attachment' --show-managed-fields=true error: exit status 1 stderr: E1013 16:52:24.921496 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:25.164255 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:25.408326 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:25.652351 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:25.897925 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:26.139650 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:26.384099 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:26.627831 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:26.872205 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:27.113456 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:27.359679 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:27.637705 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:27.882973 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:28.133521 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:28.377639 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:28.619632 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:28.865286 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:29.109654 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:29.356897 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:29.598031 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:29.837317 39172 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-disk-attachment -o json -l 'testcase=vm-disk-attachment' stderr: E1013 16:52:30.529132 39173 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:30.776072 39173 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:31.025327 39173 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:31.274482 39173 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:31.528236 39173 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:31.830300 39173 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=vm-disk-attachment", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 16:52:31.961 (8.034s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/13/25 16:52:31.961 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/13/25 16:52:31.961 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/13/25 16:52:31.961 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/13/25 16:52:31.962 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/13/25 16:52:31.962 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/13/25 16:52:31.962 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/13/25 16:52:31.962 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/13/25 16:52:31.962 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/13/25 16:52:31.962 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/13/25 16:52:31.962 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/13/25 16:52:31.962 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/13/25 16:52:31.962 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-connectivity Unexpected error: <*exec.ExitError | 0x140005ca380>: exit status 1 { ProcessState: { pid: 39174, status: 256, rusage: { Utime: { Sec: 0, Usec: 37777, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 20307, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46825472, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3668, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 37, Nvcsw: 14, Nivcsw: 646, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:63 @ 10/13/25 16:52:32.823 + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/13/25 16:52:31.963 [FAILED] kubectl create namespace head-5073ae15-end-to-end-connectivity Unexpected error: <*exec.ExitError | 0x140005ca380>: exit status 1 { ProcessState: { pid: 39174, status: 256, rusage: { Utime: { Sec: 0, Usec: 37777, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 20307, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46825472, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3668, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 37, Nvcsw: 14, Nivcsw: 646, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:63 @ 10/13/25 16:52:32.823 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/13/25 16:52:32.823 (861ms) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 16:52:32.823 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-connectivity -o yaml -l 'testcase=vm-connectivity' --show-managed-fields=true error: exit status 1 stderr: E1013 16:52:33.578850 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:33.820418 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:34.068123 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:34.309935 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:34.561110 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:34.804641 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:35.051012 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:35.301649 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:35.545186 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:35.786416 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:36.035771 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:36.278030 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:36.524608 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:36.764716 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:37.053615 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:37.300882 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:37.559272 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:37.796756 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:38.037789 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:38.279112 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:38.519185 39176 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-connectivity -o json -l 'testcase=vm-connectivity' stderr: E1013 16:52:39.432749 39177 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:39.676834 39177 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:39.919853 39177 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:40.159677 39177 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:40.403660 39177 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:40.646362 39177 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=vm-connectivity", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 16:52:40.773 (7.949s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/13/25 16:52:40.773 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/13/25 16:52:40.773 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:118 @ 10/13/25 16:52:40.773 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:129 @ 10/13/25 16:52:40.773 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:142 @ 10/13/25 16:52:40.773 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:166 @ 10/13/25 16:52:40.774 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:174 @ 10/13/25 16:52:40.774 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:179 @ 10/13/25 16:52:40.774 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:187 @ 10/13/25 16:52:40.774 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:198 @ 10/13/25 16:52:40.774 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:209 @ 10/13/25 16:52:40.774 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:222 @ 10/13/25 16:52:40.774 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:228 @ 10/13/25 16:52:40.774 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:240 @ 10/13/25 16:52:40.774 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:250 @ 10/13/25 16:52:40.774 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:258 @ 10/13/25 16:52:40.774 + + + + > Enter [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/13/25 16:52:40.774 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:73 @ 10/13/25 16:52:40.775 < Exit [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/13/25 16:52:40.775 (0s) > Enter [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/13/25 16:52:40.775 < Exit [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/13/25 16:52:40.775 (0s) > Enter [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/13/25 16:52:40.775 < Exit [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/13/25 16:52:40.775 (0s) + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-ipam Unexpected error: <*exec.ExitError | 0x1400035d660>: exit status 1 { ProcessState: { pid: 39179, status: 256, rusage: { Utime: { Sec: 0, Usec: 29575, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 11159, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47415296, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3706, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 48, Nvcsw: 24, Nivcsw: 613, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:56 @ 10/13/25 16:52:41.587 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/13/25 16:52:40.775 [FAILED] kubectl create namespace head-5073ae15-end-to-end-ipam Unexpected error: <*exec.ExitError | 0x1400035d660>: exit status 1 { ProcessState: { pid: 39179, status: 256, rusage: { Utime: { Sec: 0, Usec: 29575, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 11159, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47415296, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3706, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 48, Nvcsw: 24, Nivcsw: 613, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:56 @ 10/13/25 16:52:41.587 < Exit [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/13/25 16:52:41.587 (812ms) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 16:52:41.587 [PANICKED] Test Panicked In [AfterEach] at: /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/panic.go:262 @ 10/13/25 16:52:41.588 runtime error: invalid memory address or nil pointer dereference Full Stack Trace github.com/deckhouse/virtualization/tests/e2e.init.func8.5() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:158 +0x20 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 16:52:41.588 (0s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/13/25 16:52:41.588 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-vd-snapshots Unexpected error: <*exec.ExitError | 0x140005ca660>: exit status 1 { ProcessState: { pid: 39182, status: 256, rusage: { Utime: { Sec: 0, Usec: 29506, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 10624, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47120384, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3682, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 14, Nsignals: 48, Nvcsw: 15, Nivcsw: 631, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:69 @ 10/13/25 16:52:42.297 + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/13/25 16:52:41.588 [FAILED] kubectl create namespace head-5073ae15-end-to-end-vd-snapshots Unexpected error: <*exec.ExitError | 0x140005ca660>: exit status 1 { ProcessState: { pid: 39182, status: 256, rusage: { Utime: { Sec: 0, Usec: 29506, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 10624, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47120384, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3682, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 14, Nsignals: 48, Nvcsw: 15, Nivcsw: 631, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:69 @ 10/13/25 16:52:42.297 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/13/25 16:52:42.297 (709ms) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/13/25 16:52:42.297 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vd-snapshots -o yaml -l 'id=head-5073ae15,testcase=vd-snapshots' --show-managed-fields=true error: exit status 1 stderr: E1013 16:52:43.188792 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:43.432180 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:43.684134 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:43.930309 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:44.177580 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:44.419913 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:44.677757 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:44.921955 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:45.175400 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:45.418947 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:45.664278 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:45.907818 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:46.156927 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:46.398773 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:46.681074 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:46.938672 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:47.183434 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:47.426635 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:47.668414 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:47.911824 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:48.155142 39183 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vd-snapshots -o json -l 'id=head-5073ae15,testcase=vd-snapshots' stderr: E1013 16:52:48.855470 39185 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:49.103499 39185 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:49.353114 39185 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:49.592528 39185 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:49.838721 39185 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:50.078029 39185 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "id=head-5073ae15,testcase=vd-snapshots", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/13/25 16:52:50.205 (7.908s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:100 @ 10/13/25 16:52:50.205 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:111 @ 10/13/25 16:52:50.205 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:122 @ 10/13/25 16:52:50.205 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:133 @ 10/13/25 16:52:50.205 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:144 @ 10/13/25 16:52:50.205 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:164 @ 10/13/25 16:52:50.206 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:185 @ 10/13/25 16:52:50.206 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:229 @ 10/13/25 16:52:50.206 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:290 @ 10/13/25 16:52:50.206 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:341 @ 10/13/25 16:52:50.206 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:367 @ 10/13/25 16:52:50.206 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-label-annotation Unexpected error: <*exec.ExitError | 0x14000696460>: exit status 1 { ProcessState: { pid: 39186, status: 256, rusage: { Utime: { Sec: 0, Usec: 30680, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 15067, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46596096, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3652, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 14, Nsignals: 51, Nvcsw: 45, Nivcsw: 642, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:48 @ 10/13/25 16:52:50.72 + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/13/25 16:52:50.206 [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-label-annotation Unexpected error: <*exec.ExitError | 0x14000696460>: exit status 1 { ProcessState: { pid: 39186, status: 256, rusage: { Utime: { Sec: 0, Usec: 30680, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 15067, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46596096, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3652, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 14, Nsignals: 51, Nvcsw: 45, Nivcsw: 642, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:48 @ 10/13/25 16:52:50.72 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/13/25 16:52:50.721 (515ms) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 16:52:50.721 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-label-annotation -o yaml -l 'testcase=vm-label-annotation' --show-managed-fields=true error: exit status 1 stderr: E1013 16:52:51.347405 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:51.586637 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:51.871465 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:52.111280 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:52.353998 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:52.599880 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:52.844099 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:53.082975 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:53.325741 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:53.564522 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:53.807070 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:54.046367 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:54.293910 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:54.532946 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:54.775958 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:55.014721 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:55.254841 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:55.493091 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:55.746407 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:55.983848 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:56.222689 39187 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-label-annotation -o json -l 'testcase=vm-label-annotation' stderr: E1013 16:52:56.866066 39190 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:57.111050 39190 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:57.369721 39190 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:57.613190 39190 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:57.861164 39190 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:58.106821 39190 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=vm-label-annotation", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 16:52:58.238 (7.517s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/13/25 16:52:58.238 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/13/25 16:52:58.238 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/13/25 16:52:58.238 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/13/25 16:52:58.238 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/13/25 16:52:58.238 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/13/25 16:52:58.238 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/13/25 16:52:58.238 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/13/25 16:52:58.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/13/25 16:52:58.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/13/25 16:52:58.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/13/25 16:52:58.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/13/25 16:52:58.239 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-restore-force Unexpected error: <*exec.ExitError | 0x140005ca000>: exit status 1 { ProcessState: { pid: 39191, status: 256, rusage: { Utime: { Sec: 0, Usec: 30019, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 11785, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47857664, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3727, Majflt: 6, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 46, Nvcsw: 21, Nivcsw: 690, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:61 @ 10/13/25 16:52:58.755 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/13/25 16:52:58.239 [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-restore-force Unexpected error: <*exec.ExitError | 0x140005ca000>: exit status 1 { ProcessState: { pid: 39191, status: 256, rusage: { Utime: { Sec: 0, Usec: 30019, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 11785, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47857664, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3727, Majflt: 6, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 46, Nvcsw: 21, Nivcsw: 690, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:61 @ 10/13/25 16:52:58.755 < Exit [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/13/25 16:52:58.755 (516ms) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 16:52:58.755 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-restore-force -o yaml -l 'testcase=vm-restore-force' --show-managed-fields=true error: exit status 1 stderr: E1013 16:52:59.384401 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:59.634121 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:52:59.879698 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:00.123543 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:00.378821 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:00.626956 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:00.876392 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:01.120434 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:01.389778 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:01.634877 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:01.885235 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:02.127427 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:02.382710 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:02.626172 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:02.873618 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:03.121910 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:03.371630 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:03.622993 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:03.870888 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:04.118591 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:04.363992 39192 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-restore-force -o json -l 'testcase=vm-restore-force' stderr: E1013 16:53:05.278987 39196 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:05.515825 39196 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:05.760549 39196 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:05.997199 39196 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:06.234505 39196 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:06.488461 39196 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=vm-restore-force", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. [PANICKED] Test Panicked In [AfterEach] at: /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/panic.go:262 @ 10/13/25 16:53:06.61 runtime error: invalid memory address or nil pointer dereference Full Stack Trace github.com/deckhouse/virtualization/tests/e2e.init.func22.3() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:76 +0xe4 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 16:53:06.61 (7.855s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/13/25 16:53:06.61 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:131 @ 10/13/25 16:53:06.61 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:177 @ 10/13/25 16:53:06.611 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:341 @ 10/13/25 16:53:06.611 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:367 @ 10/13/25 16:53:06.611 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-images-creation Unexpected error: <*exec.ExitError | 0x1400065b920>: exit status 1 { ProcessState: { pid: 39197, status: 256, rusage: { Utime: { Sec: 0, Usec: 33547, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 13829, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47169536, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3686, Majflt: 4, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 14, Nsignals: 45, Nvcsw: 36, Nivcsw: 619, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:49 @ 10/13/25 16:53:07.433 + > Enter [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/13/25 16:53:06.611 [FAILED] kubectl create namespace head-5073ae15-end-to-end-images-creation Unexpected error: <*exec.ExitError | 0x1400065b920>: exit status 1 { ProcessState: { pid: 39197, status: 256, rusage: { Utime: { Sec: 0, Usec: 33547, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 13829, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47169536, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3686, Majflt: 4, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 14, Nsignals: 45, Nvcsw: 36, Nivcsw: 619, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:49 @ 10/13/25 16:53:07.433 < Exit [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/13/25 16:53:07.434 (823ms) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/13/25 16:53:07.434 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-images-creation -o yaml -l 'testcase=images-creation' --show-managed-fields=true error: exit status 1 stderr: E1013 16:53:08.219193 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:08.460769 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:08.706890 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:08.952050 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:09.195284 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:09.440041 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:09.706022 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:09.948901 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:10.192042 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:10.441178 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:10.684209 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:10.927103 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:11.173401 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:11.415555 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:11.659983 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:11.901099 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:12.232707 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:12.479939 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:12.725478 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:12.976206 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:13.218935 39198 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-images-creation -o json -l 'testcase=images-creation' stderr: E1013 16:53:13.966808 39210 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:14.213261 39210 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:14.463741 39210 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:14.708403 39210 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:14.954085 39210 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:15.241297 39210 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=images-creation", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/13/25 16:53:15.364 (7.931s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:94 @ 10/13/25 16:53:15.365 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:103 @ 10/13/25 16:53:15.365 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:114 @ 10/13/25 16:53:15.365 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:128 @ 10/13/25 16:53:15.365 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:139 @ 10/13/25 16:53:15.365 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-evacuation Unexpected error: <*exec.ExitError | 0x140002b56e0>: exit status 1 { ProcessState: { pid: 39211, status: 256, rusage: { Utime: { Sec: 0, Usec: 30900, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 12659, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47546368, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3712, Majflt: 4, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 11, Nsignals: 63, Nvcsw: 55, Nivcsw: 557, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:48 @ 10/13/25 16:53:16.259 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/13/25 16:53:15.365 [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-evacuation Unexpected error: <*exec.ExitError | 0x140002b56e0>: exit status 1 { ProcessState: { pid: 39211, status: 256, rusage: { Utime: { Sec: 0, Usec: 30900, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 12659, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47546368, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3712, Majflt: 4, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 11, Nsignals: 63, Nvcsw: 55, Nivcsw: 557, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:48 @ 10/13/25 16:53:16.259 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/13/25 16:53:16.259 (894ms) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/13/25 16:53:16.259 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-evacuation -o yaml -l 'testcase=vm-evacuation' --show-managed-fields=true error: exit status 1 stderr: E1013 16:53:16.949323 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:17.195515 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:17.437925 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:17.728452 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:17.971731 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:18.265966 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:18.507946 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:18.748065 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:18.991261 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:19.231731 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:19.481346 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:19.736925 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:19.978583 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:20.217469 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:20.463989 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:20.701909 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:21.005866 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:21.255862 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:21.502911 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:21.878396 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:22.128457 39214 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-evacuation -o json -l 'testcase=vm-evacuation' stderr: E1013 16:53:23.149276 39224 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:23.386097 39224 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:23.625543 39224 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:23.863992 39224 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:24.104358 39224 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:24.341590 39224 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=vm-evacuation", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 16:53:24.464 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 16:53:31.422 (6.958s) [FAILED] cannot delete test case resources cmd: kubectl delete virtualmachineoperations.virtualization.deckhouse.io --namespace head-5073ae15-end-to-end-vm-evacuation -l 'testcase=vm-evacuation' stderr: E1013 16:53:28.945587 39233 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:29.190281 39233 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:29.440388 39233 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:29.685784 39233 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:29.937269 39233 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:30.183150 39233 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:30.432110 39233 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:30.677347 39233 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:30.923159 39233 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:31.170796 39233 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:31.415769 39233 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Unexpected error: <*exec.ExitError | 0x1400065bd20>: exit status 1 { ProcessState: { pid: 39233, status: 256, rusage: { Utime: { Sec: 0, Usec: 98647, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 87847, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 50479104, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3892, Majflt: 4, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 33, Msgrcv: 33, Nsignals: 70, Nvcsw: 45, Nivcsw: 2524, }, }, Stderr: nil, } occurred In [AfterEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:626 @ 10/13/25 16:53:31.423 < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/13/25 16:53:31.423 (15.163s) + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-restore-safe Unexpected error: <*exec.ExitError | 0x140000ef3c0>: exit status 1 { ProcessState: { pid: 39236, status: 256, rusage: { Utime: { Sec: 0, Usec: 36584, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 18316, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47202304, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3691, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 11, Nsignals: 44, Nvcsw: 33, Nivcsw: 594, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:58 @ 10/13/25 16:53:32.335 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/13/25 16:53:31.423 [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-restore-safe Unexpected error: <*exec.ExitError | 0x140000ef3c0>: exit status 1 { ProcessState: { pid: 39236, status: 256, rusage: { Utime: { Sec: 0, Usec: 36584, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 18316, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47202304, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3691, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 11, Nsignals: 44, Nvcsw: 33, Nivcsw: 594, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:58 @ 10/13/25 16:53:32.335 < Exit [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/13/25 16:53:32.336 (912ms) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 16:53:32.336 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-restore-safe -o yaml -l 'testcase=vm-restore-safe' --show-managed-fields=true error: exit status 1 stderr: E1013 16:53:33.017598 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:33.258517 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:33.507181 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:33.747714 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:33.993614 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:34.243262 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:34.484186 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:34.720692 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:34.964357 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:35.204079 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:35.448666 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:35.689871 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:35.937261 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:36.180262 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:36.428250 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:36.680826 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:36.925111 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:37.165522 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:37.406088 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:37.645687 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:37.885054 39237 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-restore-safe -o json -l 'testcase=vm-restore-safe' stderr: E1013 16:53:38.540549 39241 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:38.787168 39241 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:39.038150 39241 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:39.282281 39241 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:39.532298 39241 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:39.777185 39241 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=vm-restore-safe", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. [PANICKED] Test Panicked In [AfterEach] at: /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/panic.go:262 @ 10/13/25 16:53:39.909 runtime error: invalid memory address or nil pointer dereference Full Stack Trace github.com/deckhouse/virtualization/tests/e2e.init.func23.3() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:70 +0xe4 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 16:53:39.909 (7.573s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/13/25 16:53:39.909 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:115 @ 10/13/25 16:53:39.909 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:158 @ 10/13/25 16:53:39.909 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:348 @ 10/13/25 16:53:39.909 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:374 @ 10/13/25 16:53:39.91 + + + + > Enter [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/13/25 16:53:39.91 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:51 @ 10/13/25 16:53:39.91 < Exit [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/13/25 16:53:39.91 (0s) > Enter [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/13/25 16:53:39.91 < Exit [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/13/25 16:53:39.91 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:82 @ 10/13/25 16:53:39.91 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:94 @ 10/13/25 16:53:39.91 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:106 @ 10/13/25 16:53:39.91 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:116 @ 10/13/25 16:53:39.91 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:132 @ 10/13/25 16:53:39.911 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:151 @ 10/13/25 16:53:39.911 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:157 @ 10/13/25 16:53:39.911 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:189 @ 10/13/25 16:53:39.911 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:216 @ 10/13/25 16:53:39.911 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:252 @ 10/13/25 16:53:39.911 + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.911 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 16:53:39.911 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.911 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.912 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 16:53:39.912 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.912 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.912 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 16:53:39.912 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.912 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.912 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 16:53:39.912 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.912 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.912 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 16:53:39.912 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.913 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.913 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 16:53:39.913 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.913 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.913 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 16:53:39.913 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.913 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.913 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 16:53:39.913 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.913 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.913 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 16:53:39.913 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.913 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.914 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 16:53:39.914 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.914 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.914 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 16:53:39.914 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.914 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.914 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 16:53:39.914 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 16:53:39.914 (0s) + + + + > Enter [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/13/25 16:53:39.914 [SKIPPED] Module SDN is disabled. Skipping all tests for module SDN. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:50 @ 10/13/25 16:53:40.412 < Exit [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/13/25 16:53:40.413 (499ms) > Enter [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/13/25 16:53:40.413 < Exit [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/13/25 16:53:40.413 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/13/25 16:53:40.413 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/13/25 16:53:40.413 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:111 @ 10/13/25 16:53:40.414 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:125 @ 10/13/25 16:53:40.414 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:141 @ 10/13/25 16:53:40.414 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:156 @ 10/13/25 16:53:40.414 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:169 @ 10/13/25 16:53:40.414 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-configuration Unexpected error: <*exec.ExitError | 0x140005ca000>: exit status 1 { ProcessState: { pid: 39242, status: 256, rusage: { Utime: { Sec: 0, Usec: 41985, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 21101, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47267840, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3704, Majflt: 4, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 14, Nsignals: 62, Nvcsw: 41, Nivcsw: 690, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:55 @ 10/13/25 16:53:41.4 + > Enter [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/13/25 16:53:40.415 [FAILED] kubectl create namespace head-5073ae15-end-to-end-vm-configuration Unexpected error: <*exec.ExitError | 0x140005ca000>: exit status 1 { ProcessState: { pid: 39242, status: 256, rusage: { Utime: { Sec: 0, Usec: 41985, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 21101, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47267840, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3704, Majflt: 4, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 14, Nsignals: 62, Nvcsw: 41, Nivcsw: 690, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:55 @ 10/13/25 16:53:41.4 < Exit [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/13/25 16:53:41.401 (986ms) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 16:53:41.401 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-vm-configuration -o yaml -l 'testcase=vm-configuration' --show-managed-fields=true error: exit status 1 stderr: E1013 16:53:42.079648 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:42.330290 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:42.580118 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:42.820004 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:43.066867 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:43.306584 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:43.555232 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:43.796767 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:44.044304 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:44.284490 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:44.531285 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:44.780028 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:45.028258 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:45.277089 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:45.522234 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:45.797444 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:46.051816 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:46.299162 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:46.540865 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:46.783023 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:47.024895 39245 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-vm-configuration -o json -l 'testcase=vm-configuration' stderr: E1013 16:53:47.970129 39247 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:48.221934 39247 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:48.474446 39247 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:48.721394 39247 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:48.972106 39247 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:49.219308 39247 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=vm-configuration", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 16:53:49.351 (7.951s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/13/25 16:53:49.352 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/13/25 16:53:49.352 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:109 @ 10/13/25 16:53:49.352 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:123 @ 10/13/25 16:53:49.352 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:147 @ 10/13/25 16:53:49.352 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:161 @ 10/13/25 16:53:49.352 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:183 @ 10/13/25 16:53:49.352 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:202 @ 10/13/25 16:53:49.352 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:226 @ 10/13/25 16:53:49.352 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:240 @ 10/13/25 16:53:49.353 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:250 @ 10/13/25 16:53:49.353 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:265 @ 10/13/25 16:53:49.353 + + + [FAILED] kubectl create namespace head-5073ae15-end-to-end-complex-test Unexpected error: <*exec.ExitError | 0x1400035c920>: exit status 1 { ProcessState: { pid: 39248, status: 256, rusage: { Utime: { Sec: 0, Usec: 41988, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 20427, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47333376, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3702, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 14, Nsignals: 54, Nvcsw: 36, Nivcsw: 629, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:58 @ 10/13/25 16:53:50.004 + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/13/25 16:53:49.353 [FAILED] kubectl create namespace head-5073ae15-end-to-end-complex-test Unexpected error: <*exec.ExitError | 0x1400035c920>: exit status 1 { ProcessState: { pid: 39248, status: 256, rusage: { Utime: { Sec: 0, Usec: 41988, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 20427, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47333376, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3702, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 14, Nsignals: 54, Nvcsw: 36, Nivcsw: 629, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:58 @ 10/13/25 16:53:50.004 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/13/25 16:53:50.005 (652ms) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 16:53:50.005 Get resources error: cmd: kubectl get virtualization,cvi,vmc,intvirt,pod,volumesnapshot --namespace head-5073ae15-end-to-end-complex-test -o yaml -l 'testcase=complex-test' --show-managed-fields=true error: exit status 1 stderr: E1013 16:53:50.997421 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:51.242971 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:51.491952 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:51.738906 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:51.987719 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:52.234365 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:52.483076 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:52.729599 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:52.978869 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:53.224766 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:53.483342 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:53.738690 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:53.990671 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:54.237482 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:54.494238 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:54.744157 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:54.997192 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:55.256659 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:55.502600 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:55.748457 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:55.994984 39249 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Failed to get PodList: cmd: kubectl get pod --namespace head-5073ae15-end-to-end-complex-test -o json -l 'testcase=complex-test' stderr: E1013 16:53:57.112411 39251 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:57.361116 39251 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:57.624926 39251 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:57.871113 39251 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:58.127609 39251 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1013 16:53:58.373939 39251 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): Unable to find "/v1, Resource=pods" that match label selector "testcase=complex-test", field selector "": the server could not find the requested resource (get pods) The list of pods is empty; nothing to dump. < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 16:53:58.507 (8.502s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/13/25 16:53:58.507 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/13/25 16:53:58.507 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/13/25 16:53:58.507 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/13/25 16:53:58.507 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/13/25 16:53:58.507 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:141 @ 10/13/25 16:53:58.507 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:152 @ 10/13/25 16:53:58.507 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:162 @ 10/13/25 16:53:58.508 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:173 @ 10/13/25 16:53:58.508 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:184 @ 10/13/25 16:53:58.508 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:196 @ 10/13/25 16:53:58.508 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:220 @ 10/13/25 16:53:58.508 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:245 @ 10/13/25 16:53:58.508 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:268 @ 10/13/25 16:53:58.508 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:280 @ 10/13/25 16:53:58.508 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:299 @ 10/13/25 16:53:58.508 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:316 @ 10/13/25 16:53:58.508 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:341 @ 10/13/25 16:53:58.509 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:354 @ 10/13/25 16:53:58.509 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:358 @ 10/13/25 16:53:58.509 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:375 @ 10/13/25 16:53:58.509 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:388 @ 10/13/25 16:53:58.509 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:405 @ 10/13/25 16:53:58.509 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:444 @ 10/13/25 16:53:58.509 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:479 @ 10/13/25 16:53:58.509 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:497 @ 10/13/25 16:53:58.509 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:512 @ 10/13/25 16:53:58.509 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:528 @ 10/13/25 16:53:58.509 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:545 @ 10/13/25 16:53:58.509 + + + [FAILED] Expected success, but got an error: <*errors.StatusError | 0x1400044a640>: the server could not find the requested resource (get pods) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "the server could not find the requested resource (get pods)", Reason: "NotFound", Details: { Name: "", Group: "", Kind: "pods", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>404 Not Found</title></head>\r\n<body>\r\n<center><h1>404 Not Found</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 404, }, } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/13/25 16:53:58.645 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 16:53:58.509 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 16:53:58.509 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 16:53:58.509 [FAILED] Expected success, but got an error: <*errors.StatusError | 0x1400044a640>: the server could not find the requested resource (get pods) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "the server could not find the requested resource (get pods)", Reason: "NotFound", Details: { Name: "", Group: "", Kind: "pods", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>404 Not Found</title></head>\r\n<body>\r\n<center><h1>404 Not Found</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 404, }, } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/13/25 16:53:58.645 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 16:53:58.645 (136ms) + + + [FAILED] Expected success, but got an error: <*errors.StatusError | 0x14000173400>: the server could not find the requested resource (get projects.deckhouse.io) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "the server could not find the requested resource (get projects.deckhouse.io)", Reason: "NotFound", Details: { Name: "", Group: "deckhouse.io", Kind: "projects", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>404 Not Found</title></head>\r\n<body>\r\n<center><h1>404 Not Found</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 404, }, } In [DeferCleanup (Suite)] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:177 @ 10/13/25 16:53:58.766 + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/13/25 16:53:58.646 [FAILED] Expected success, but got an error: <*errors.StatusError | 0x14000173400>: the server could not find the requested resource (get projects.deckhouse.io) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "the server could not find the requested resource (get projects.deckhouse.io)", Reason: "NotFound", Details: { Name: "", Group: "deckhouse.io", Kind: "projects", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>404 Not Found</title></head>\r\n<body>\r\n<center><h1>404 Not Found</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 404, }, } In [DeferCleanup (Suite)] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:177 @ 10/13/25 16:53:58.766 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/13/25 16:53:58.767 (121ms) + + + \ No newline at end of file diff --git a/artifacts/clean-sds-20251013-173506-29678-rerun/junit.xml b/artifacts/clean-sds-20251013-173506-29678-rerun/junit.xml new file mode 100644 index 0000000000..d8a633a96f --- /dev/null +++ b/artifacts/clean-sds-20251013-173506-29678-rerun/junit.xml @@ -0,0 +1,2 @@ + + \ No newline at end of file diff --git a/artifacts/clean-sds-20251013-173506-29678/junit.xml b/artifacts/clean-sds-20251013-173506-29678/junit.xml new file mode 100644 index 0000000000..d8a633a96f --- /dev/null +++ b/artifacts/clean-sds-20251013-173506-29678/junit.xml @@ -0,0 +1,2 @@ + + \ No newline at end of file diff --git a/artifacts/clean-sds-20251013-183813-30975/junit.xml b/artifacts/clean-sds-20251013-183813-30975/junit.xml new file mode 100644 index 0000000000..47c7c72fba --- /dev/null +++ b/artifacts/clean-sds-20251013-183813-30975/junit.xml @@ -0,0 +1,856 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 19:04:54.06 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 19:04:58.653 (4.592s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 19:04:58.653 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/13/25 19:04:58.653 (0s) + + + > Enter [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/13/25 19:04:58.654 < Exit [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/13/25 19:04:58.654 (0s) > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:04:58.654 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:04:58.654 (0s) > Enter [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/13/25 19:04:58.654 < Exit [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/13/25 19:05:00.011 (1.357s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:05:00.011 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:05:00.012 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:05:00.012 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:05:00.012 (0s) > Enter [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/13/25 19:05:00.012 STEP: Project should be deployed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:73 @ 10/13/25 19:05:00.012 < Exit [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/13/25 19:05:02.722 (2.71s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:05:02.722 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:05:02.722 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:05:02.722 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:05:02.722 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/13/25 19:05:02.722 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/13/25 19:05:05.705 (2.983s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:05:05.705 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:05:05.705 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:05:05.705 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:05:05.706 (0s) > Enter [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/13/25 19:05:05.706 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/13/25 19:05:05.706 < Exit [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/13/25 19:05:12.079 (6.373s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:05:12.079 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:05:12.079 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-cvi --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-cvi\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/13/25 19:21:54.685 + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:05:12.08 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/13/25 19:05:12.08 (0s) > Enter [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/13/25 19:05:12.08 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/13/25 19:05:12.08 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-cvi --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-cvi\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/13/25 19:21:54.685 < Exit [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/13/25 19:21:54.686 (16m42.586s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:21:54.686 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/13/25 19:22:02.87 (8.184s) > Enter [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/13/25 19:22:02.87 STEP: Delete manifests - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:50 @ 10/13/25 19:22:02.87 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 19:22:02.87 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 19:22:05.022 (2.152s) < Exit [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/13/25 19:22:05.022 (2.152s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/13/25 19:22:05.023 + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.023 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 19:22:05.023 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.023 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.023 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 19:22:05.023 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.023 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.023 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 19:22:05.024 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.024 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.024 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 19:22:05.024 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.024 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.024 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 19:22:05.024 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.024 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.024 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 19:22:05.024 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.024 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.024 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 19:22:05.024 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.025 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.025 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 19:22:05.025 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.025 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.025 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/13/25 19:22:05.025 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/13/25 19:22:05.025 (0s) + + + + > Enter [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/13/25 19:22:05.025 [SKIPPED] Module SDN is disabled. Skipping all tests for module SDN. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:50 @ 10/13/25 19:22:05.142 < Exit [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/13/25 19:22:05.142 (117ms) > Enter [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/13/25 19:22:05.142 < Exit [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/13/25 19:22:05.142 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/13/25 19:22:05.142 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/13/25 19:22:05.143 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:111 @ 10/13/25 19:22:05.143 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:125 @ 10/13/25 19:22:05.143 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:141 @ 10/13/25 19:22:05.143 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:156 @ 10/13/25 19:22:05.143 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:169 @ 10/13/25 19:22:05.143 + + + > Enter [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/13/25 19:22:05.144 < Exit [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/13/25 19:22:07.118 (1.974s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/13/25 19:22:07.118 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/13/25 19:22:12.805 (5.688s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 19:22:12.805 < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 19:22:12.805 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:77 @ 10/13/25 19:38:55.327 + > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/13/25 19:22:12.805 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:76 @ 10/13/25 19:22:12.806 [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:77 @ 10/13/25 19:38:55.327 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/13/25 19:38:55.328 (16m42.477s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 19:38:55.328 < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/13/25 19:39:06.251 (10.923s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:86 @ 10/13/25 19:39:06.251 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:100 @ 10/13/25 19:39:06.251 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:116 @ 10/13/25 19:39:06.251 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:131 @ 10/13/25 19:39:06.251 + + + + > Enter [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/13/25 19:39:06.251 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:73 @ 10/13/25 19:39:06.252 < Exit [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/13/25 19:39:06.252 (0s) > Enter [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/13/25 19:39:06.252 < Exit [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/13/25 19:39:06.252 (0s) > Enter [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/13/25 19:39:06.252 < Exit [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/13/25 19:39:06.252 (0s) + + + + > Enter [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/13/25 19:39:06.252 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:41 @ 10/13/25 19:39:06.252 < Exit [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/13/25 19:39:06.252 (0s) > Enter [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/13/25 19:39:06.252 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 19:39:06.252 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 19:39:10.327 (4.075s) < Exit [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/13/25 19:39:10.327 (4.075s) + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/13/25 19:39:11.847 + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/13/25 19:39:10.328 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/13/25 19:39:11.847 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/13/25 19:39:11.847 (1.519s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/13/25 19:39:11.847 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/13/25 19:39:17.685 (5.838s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:100 @ 10/13/25 19:39:17.686 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:111 @ 10/13/25 19:39:17.686 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:122 @ 10/13/25 19:39:17.686 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:133 @ 10/13/25 19:39:17.686 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:144 @ 10/13/25 19:39:17.686 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:164 @ 10/13/25 19:39:17.686 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:185 @ 10/13/25 19:39:17.686 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:229 @ 10/13/25 19:39:17.686 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:290 @ 10/13/25 19:39:17.686 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:341 @ 10/13/25 19:39:17.686 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:367 @ 10/13/25 19:39:17.686 + + + [FAILED] Warning: resource namespaces/head-5073ae15-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "linstor-thin-r2" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x140004c1900>: exit status 1 { ProcessState: { pid: 48174, status: 256, rusage: { Utime: { Sec: 0, Usec: 234268, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 41764, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 70713344, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5126, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 48, Msgrcv: 62, Nsignals: 278, Nvcsw: 202, Nivcsw: 2888, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/13/25 19:39:24.91 + > Enter [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/13/25 19:39:17.686 < Exit [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/13/25 19:39:21.053 (3.367s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/13/25 19:39:21.053 [FAILED] Warning: resource namespaces/head-5073ae15-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "linstor-thin-r2" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x140004c1900>: exit status 1 { ProcessState: { pid: 48174, status: 256, rusage: { Utime: { Sec: 0, Usec: 234268, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 41764, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 70713344, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5126, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 48, Msgrcv: 62, Nsignals: 278, Nvcsw: 202, Nivcsw: 2888, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/13/25 19:39:24.91 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/13/25 19:39:24.91 (3.857s) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/13/25 19:39:24.91 The list of pods is empty; nothing to dump. < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/13/25 19:39:30.908 (5.998s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:112 @ 10/13/25 19:39:30.908 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:147 @ 10/13/25 19:39:30.908 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:190 @ 10/13/25 19:39:30.909 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:196 @ 10/13/25 19:39:30.909 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:204 @ 10/13/25 19:39:30.909 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:224 @ 10/13/25 19:39:30.909 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:236 @ 10/13/25 19:39:30.909 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:258 @ 10/13/25 19:39:30.909 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:284 @ 10/13/25 19:39:30.909 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:294 @ 10/13/25 19:39:30.909 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:309 @ 10/13/25 19:39:30.909 + + + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/13/25 19:39:30.909 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/13/25 19:39:32.384 (1.475s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/13/25 19:39:32.384 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/13/25 19:39:36.208 (3.824s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 19:39:36.208 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 19:39:36.208 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/13/25 19:39:36.208 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:97 @ 10/13/25 19:39:36.208 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/13/25 19:39:50.123 (13.914s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 19:39:50.123 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 19:39:50.123 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-b --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-a --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 @ 10/13/25 19:56:32.295 + > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/13/25 19:39:50.123 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 @ 10/13/25 19:39:50.123 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-b --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-connectivity-a --namespace head-5073ae15-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-connectivity-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 @ 10/13/25 19:56:32.295 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/13/25 19:56:32.295 (16m42.136s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 19:56:32.296 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/13/25 19:56:43.18 (10.884s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:118 @ 10/13/25 19:56:43.18 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:129 @ 10/13/25 19:56:43.18 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:142 @ 10/13/25 19:56:43.18 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:166 @ 10/13/25 19:56:43.18 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:174 @ 10/13/25 19:56:43.181 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:179 @ 10/13/25 19:56:43.181 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:187 @ 10/13/25 19:56:43.181 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:198 @ 10/13/25 19:56:43.181 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:209 @ 10/13/25 19:56:43.181 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:222 @ 10/13/25 19:56:43.181 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:228 @ 10/13/25 19:56:43.181 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:240 @ 10/13/25 19:56:43.181 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:250 @ 10/13/25 19:56:43.181 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:258 @ 10/13/25 19:56:43.181 + + + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/13/25 19:56:43.182 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/13/25 19:56:45.199 (2.018s) > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:56:45.199 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:56:45.199 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/13/25 19:56:45.199 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/13/25 19:56:48.096 (2.896s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:56:48.096 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:56:48.096 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:56:48.096 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:56:48.096 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/13/25 19:56:48.096 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:75 @ 10/13/25 19:56:48.096 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/13/25 19:56:53.63 (5.534s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:56:53.63 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 19:56:53.63 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-label-annotation --namespace head-5073ae15-end-to-end-vm-label-annotation --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-label-annotation\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 @ 10/13/25 20:13:35.839 + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:56:53.67 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/13/25 19:56:53.67 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/13/25 19:56:53.67 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 @ 10/13/25 19:56:53.67 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-label-annotation --namespace head-5073ae15-end-to-end-vm-label-annotation --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-label-annotation\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 @ 10/13/25 20:13:35.839 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/13/25 20:13:35.84 (16m42.142s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 20:13:35.84 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/13/25 20:13:44.541 (8.701s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/13/25 20:13:44.541 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/13/25 20:13:44.541 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/13/25 20:13:44.542 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/13/25 20:13:44.542 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/13/25 20:13:44.542 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/13/25 20:13:44.542 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/13/25 20:13:44.542 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/13/25 20:13:44.542 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/13/25 20:13:44.542 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/13/25 20:13:44.542 + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.542 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 20:13:44.543 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.543 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.543 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 20:13:44.543 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.543 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.543 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 20:13:44.543 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.543 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.543 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 20:13:44.543 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.543 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.543 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 20:13:44.543 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.543 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.544 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 20:13:44.544 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.544 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.544 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 20:13:44.544 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.544 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.544 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 20:13:44.544 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.544 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.544 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 20:13:44.544 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.544 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.545 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 20:13:44.545 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.545 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.545 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 20:13:44.545 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.545 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.545 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/13/25 20:13:44.545 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/13/25 20:13:44.545 (0s) + + + + > Enter [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/13/25 20:13:44.546 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:51 @ 10/13/25 20:13:44.546 < Exit [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/13/25 20:13:44.546 (0s) > Enter [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/13/25 20:13:44.546 < Exit [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/13/25 20:13:44.546 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:82 @ 10/13/25 20:13:44.546 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:94 @ 10/13/25 20:13:44.546 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:106 @ 10/13/25 20:13:44.546 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:116 @ 10/13/25 20:13:44.546 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:132 @ 10/13/25 20:13:44.546 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:151 @ 10/13/25 20:13:44.546 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:157 @ 10/13/25 20:13:44.546 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:189 @ 10/13/25 20:13:44.546 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:216 @ 10/13/25 20:13:44.546 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:252 @ 10/13/25 20:13:44.546 + + + > Enter [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/13/25 20:13:44.546 < Exit [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/13/25 20:13:46.023 (1.477s) > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 20:13:46.023 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 20:13:46.023 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/13/25 20:13:46.023 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/13/25 20:13:47.756 (1.732s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 20:13:47.756 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 20:13:47.756 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root --namespace head-5073ae15-end-to-end-vm-versions --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:69 @ 10/13/25 20:30:29.898 + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 20:13:47.756 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/13/25 20:13:47.756 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/13/25 20:13:47.756 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:68 @ 10/13/25 20:13:47.756 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root --namespace head-5073ae15-end-to-end-vm-versions --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:69 @ 10/13/25 20:30:29.898 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/13/25 20:30:29.898 (16m42.148s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 20:30:29.898 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/13/25 20:30:38.596 (8.698s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/13/25 20:30:38.597 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/13/25 20:30:38.597 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/13/25 20:30:38.597 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/13/25 20:30:38.597 + + + > Enter [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/13/25 20:30:38.597 < Exit [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/13/25 20:30:40.569 (1.972s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/13/25 20:30:40.569 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/13/25 20:30:45.675 (5.106s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 20:30:45.675 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 20:30:45.675 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/13/25 20:30:45.676 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:89 @ 10/13/25 20:30:45.676 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/13/25 20:30:49.68 (4.005s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 20:30:49.68 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 20:30:49.68 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:4>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-manual-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-manual-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:100 @ 10/13/25 20:47:31.8 + > Enter [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/13/25 20:30:49.681 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:4>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-manual-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-manual-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:100 @ 10/13/25 20:47:31.8 < Exit [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/13/25 20:47:31.801 (16m42.157s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 20:47:31.801 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/13/25 20:47:42.72 (10.92s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:109 @ 10/13/25 20:47:42.721 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:123 @ 10/13/25 20:47:42.721 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:147 @ 10/13/25 20:47:42.721 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:161 @ 10/13/25 20:47:42.721 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:183 @ 10/13/25 20:47:42.721 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:202 @ 10/13/25 20:47:42.721 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:226 @ 10/13/25 20:47:42.721 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:240 @ 10/13/25 20:47:42.721 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:250 @ 10/13/25 20:47:42.721 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:265 @ 10/13/25 20:47:42.721 + + + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/13/25 20:47:42.721 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/13/25 20:47:44.709 (1.987s) > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 20:47:44.709 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 20:47:44.709 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/13/25 20:47:44.709 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/13/25 20:47:48.221 (3.512s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 20:47:48.221 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 20:47:48.221 (0s) + + + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 20:47:48.221 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 20:47:48.221 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/13/25 20:47:48.221 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:88 @ 10/13/25 20:47:48.222 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/13/25 20:47:53.014 (4.792s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 20:47:53.014 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 20:47:53.014 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-with-hotplug-standalone --namespace head-5073ae15-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/13/25 21:04:34.701 + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 20:47:53.014 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/13/25 20:47:53.014 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/13/25 20:47:53.014 STEP: VDs with consumers should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:99 @ 10/13/25 20:47:53.014 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-with-hotplug-standalone --namespace head-5073ae15-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/13/25 21:04:34.701 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/13/25 21:04:34.703 (16m41.694s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 21:04:34.703 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/13/25 21:04:43.131 (8.428s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/13/25 21:04:43.131 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/13/25 21:04:43.131 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/13/25 21:04:43.131 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/13/25 21:04:43.131 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/13/25 21:04:43.131 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/13/25 21:04:43.131 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/13/25 21:04:43.131 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/13/25 21:04:43.131 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/13/25 21:04:43.131 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/13/25 21:04:43.131 + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:51 @ 10/13/25 21:04:45.067 + > Enter [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/13/25 21:04:43.131 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:51 @ 10/13/25 21:04:45.067 < Exit [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/13/25 21:04:45.067 (1.936s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/13/25 21:04:45.067 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/13/25 21:04:50.684 (5.616s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:94 @ 10/13/25 21:04:50.684 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:103 @ 10/13/25 21:04:50.684 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:114 @ 10/13/25 21:04:50.684 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:128 @ 10/13/25 21:04:50.684 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:139 @ 10/13/25 21:04:50.684 + + + > Enter [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/13/25 21:04:50.685 < Exit [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/13/25 21:04:52.673 (1.988s) > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 21:04:52.673 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 21:04:52.673 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/13/25 21:04:52.673 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/13/25 21:04:58.966 (6.293s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 21:04:58.966 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 21:04:58.967 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-always-on-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1013 21:19:51.658352 50814 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-always-on-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1013 21:19:31.515226 50813 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 @ 10/13/25 21:25:34.113 + > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 21:04:58.967 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/13/25 21:04:58.967 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/13/25 21:04:58.967 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/13/25 21:04:58.967 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/13/25 21:25:34.113 (16m42.085s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-always-on-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1013 21:19:51.658352 50814 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-always-on-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1013 21:19:31.515226 50813 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 @ 10/13/25 21:25:34.113 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/13/25 21:25:34.115 (16m42.086s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 21:25:34.115 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/13/25 21:25:44.89 (10.775s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:131 @ 10/13/25 21:25:44.89 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:177 @ 10/13/25 21:25:44.891 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:341 @ 10/13/25 21:25:44.891 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:367 @ 10/13/25 21:25:44.891 + + + > Enter [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/13/25 21:25:44.891 < Exit [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/13/25 21:25:46.867 (1.976s) > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 21:25:46.867 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 21:25:46.867 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/13/25 21:25:46.867 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/13/25 21:25:53.505 (6.638s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 21:25:53.505 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 21:25:53.505 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 21:25:53.506 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 21:25:53.506 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/13/25 21:25:53.506 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:91 @ 10/13/25 21:25:53.506 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/13/25 21:25:58.052 (4.546s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 21:25:58.052 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 21:25:58.052 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 21:25:58.052 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 21:25:58.052 (0s) > Enter [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/13/25 21:25:58.052 STEP: VDs should be in WaitForFirstConsumer phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:102 @ 10/13/25 21:25:58.052 < Exit [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/13/25 21:26:03.509 (5.457s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 21:26:03.509 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 21:26:03.509 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-existing-vmclass --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-existing-vmclass\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:117 @ 10/13/25 21:42:45.635 + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 21:26:03.509 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/13/25 21:26:03.509 (0s) > Enter [It] checks VDs phases with map[vm:existing-vmclass] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/13/25 21:26:03.509 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:116 @ 10/13/25 21:26:03.509 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-existing-vmclass --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-existing-vmclass\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:117 @ 10/13/25 21:42:45.635 < Exit [It] checks VDs phases with map[vm:existing-vmclass] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/13/25 21:42:45.635 (16m42.121s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 21:42:45.635 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/13/25 21:42:54.344 (8.709s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:126 @ 10/13/25 21:42:54.344 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:140 @ 10/13/25 21:42:54.344 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:152 @ 10/13/25 21:42:54.344 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:157 @ 10/13/25 21:42:54.344 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:163 @ 10/13/25 21:42:54.344 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:176 @ 10/13/25 21:42:54.344 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:181 @ 10/13/25 21:42:54.344 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:187 @ 10/13/25 21:42:54.345 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:202 @ 10/13/25 21:42:54.345 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:216 @ 10/13/25 21:42:54.345 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:240 @ 10/13/25 21:42:54.345 + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1013 21:54:47.247600 51866 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1013 21:54:53.074364 51867 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 @ 10/13/25 22:01:46.22 + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/13/25 21:42:54.345 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/13/25 21:42:56.316 (1.971s) > Enter [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/13/25 21:42:56.316 < Exit [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/13/25 21:43:01.159 (4.843s) > Enter [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/13/25 21:43:01.159 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 @ 10/13/25 21:43:01.159 [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1013 21:54:47.247600 51866 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1013 21:54:53.074364 51867 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 @ 10/13/25 22:01:46.22 < Exit [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/13/25 22:01:46.222 (16m42.122s) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/13/25 22:01:46.222 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 22:01:57.646 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/13/25 22:02:02.535 (4.889s) < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/13/25 22:02:02.535 (16.313s) + + + > Enter [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/13/25 22:02:02.536 < Exit [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/13/25 22:02:05.808 (3.272s) > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 22:02:05.808 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 22:02:05.808 (0s) > Enter [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/13/25 22:02:05.808 STEP: Create a vmip automatically and check its binding with a lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:81 @ 10/13/25 22:02:05.808 STEP: Remove label from the lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:87 @ 10/13/25 22:02:06.729 STEP: Wait for the label to be restored by the controller - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:96 @ 10/13/25 22:02:06.871 < Exit [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/13/25 22:02:07.137 (1.329s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 22:02:07.137 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 22:02:07.137 (0s) + + + > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 22:02:07.137 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/13/25 22:02:07.137 (0s) > Enter [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/13/25 22:02:07.137 STEP: Create an intermediate vmip automatically to allocate a new ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:109 @ 10/13/25 22:02:07.137 STEP: Delete the intermediate vmip automatically and check that the lease is released - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:115 @ 10/13/25 22:02:07.527 STEP: Reuse the released lease with a static vmip - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:124 @ 10/13/25 22:02:07.761 STEP: Delete the static vmip and lease, then create another static vmip with this ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:132 @ 10/13/25 22:02:08.129 < Exit [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/13/25 22:02:08.735 (1.598s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 22:02:08.735 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/13/25 22:02:08.735 (0s) + + + > Enter [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/13/25 22:02:08.735 < Exit [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/13/25 22:02:10.65 (1.914s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/13/25 22:02:10.65 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/13/25 22:02:20.042 (9.392s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 22:02:20.042 < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 22:02:20.042 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:6, cap:6>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-selector --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-selector\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-c-and-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-c-and-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-d --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-d\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-b-not-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-b-not-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-affinity --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-affinity\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-a-not-b --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-a-not-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:103 @ 10/13/25 22:19:13.109 + > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/13/25 22:02:20.042 STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/13/25 22:02:20.042 END STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/13/25 22:02:28.249 (8.207s) STEP: `VirtualMachineClasses` should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:95 @ 10/13/25 22:02:28.249 END STEP: `VirtualMachineClasses` should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:95 @ 10/13/25 22:02:30.998 (2.749s) STEP: `VirtualDisks` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:102 @ 10/13/25 22:02:30.998 END STEP: `VirtualDisks` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:102 @ 10/13/25 22:19:13.109 (16m42.142s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:6, cap:6>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-selector --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-selector\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-c-and-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-c-and-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-d --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-d\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-b-not-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-b-not-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-affinity --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-affinity\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-a-not-b --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-a-not-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:103 @ 10/13/25 22:19:13.109 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/13/25 22:19:13.11 (16m53.099s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 22:19:13.11 < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/13/25 22:19:33.55 (20.44s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:120 @ 10/13/25 22:19:33.551 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:283 @ 10/13/25 22:19:33.551 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:372 @ 10/13/25 22:19:33.551 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:467 @ 10/13/25 22:19:33.551 + + + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/13/25 22:19:33.551 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/13/25 22:19:35.499 (1.948s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:62 @ 10/13/25 22:19:35.499 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:62 @ 10/13/25 22:20:02.74 (27.241s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 22:20:02.74 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 22:20:02.74 (0s) + + + > Enter [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/13/25 22:20:02.74 < Exit [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/13/25 22:20:04.071 (1.331s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 22:20:04.071 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 22:20:04.071 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/13/25 22:20:04.071 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:102 @ 10/13/25 22:20:04.071 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/13/25 22:20:12.257 (8.186s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 22:20:12.258 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 22:20:12.258 (0s) + + + > Enter [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/13/25 22:20:12.258 STEP: CVIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:113 @ 10/13/25 22:20:12.258 < Exit [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/13/25 22:20:15.121 (2.863s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 22:20:15.121 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 22:20:15.121 (0s) + + + > Enter [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/13/25 22:20:15.122 STEP: VMClasses should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:124 @ 10/13/25 22:20:15.122 < Exit [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/13/25 22:20:17.887 (2.765s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 22:20:17.887 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 22:20:17.887 (0s) + + + [FAILED] Timed out after 344.829s. Expected success, but got an error: <*errors.errorString | 0x140001129d0>: cannot patch VMIP "head-5073ae15-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip { s: "cannot patch VMIP \"head-5073ae15-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:138 @ 10/13/25 22:26:02.716 + > Enter [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/13/25 22:20:17.887 [FAILED] Timed out after 344.829s. Expected success, but got an error: <*errors.errorString | 0x140001129d0>: cannot patch VMIP "head-5073ae15-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip { s: "cannot patch VMIP \"head-5073ae15-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:138 @ 10/13/25 22:26:02.716 < Exit [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/13/25 22:26:02.718 (5m44.831s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 22:26:02.718 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/13/25 22:26:35.545 (32.837s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:141 @ 10/13/25 22:26:35.546 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:152 @ 10/13/25 22:26:35.546 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:162 @ 10/13/25 22:26:35.546 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:173 @ 10/13/25 22:26:35.546 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:184 @ 10/13/25 22:26:35.546 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:196 @ 10/13/25 22:26:35.546 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:220 @ 10/13/25 22:26:35.546 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:245 @ 10/13/25 22:26:35.546 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:268 @ 10/13/25 22:26:35.546 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:280 @ 10/13/25 22:26:35.547 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:299 @ 10/13/25 22:26:35.547 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:316 @ 10/13/25 22:26:35.547 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:341 @ 10/13/25 22:26:35.547 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:354 @ 10/13/25 22:26:35.547 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:358 @ 10/13/25 22:26:35.547 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:375 @ 10/13/25 22:26:35.547 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:388 @ 10/13/25 22:26:35.547 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:405 @ 10/13/25 22:26:35.547 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:444 @ 10/13/25 22:26:35.548 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:479 @ 10/13/25 22:26:35.548 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:497 @ 10/13/25 22:26:35.548 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:512 @ 10/13/25 22:26:35.548 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:528 @ 10/13/25 22:26:35.548 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:545 @ 10/13/25 22:26:35.548 + + + > Enter [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/13/25 22:26:35.548 < Exit [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/13/25 22:26:37.063 (1.516s) > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/13/25 22:26:37.063 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/13/25 22:26:37.063 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/13/25 22:26:37.063 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/13/25 22:26:40.976 (3.915s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 22:26:40.976 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 22:26:40.976 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-safe --namespace head-5073ae15-end-to-end-vm-restore-safe --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-safe\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:107 @ 10/13/25 22:43:23.135 + > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/13/25 22:26:40.976 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/13/25 22:26:40.976 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/13/25 22:26:40.976 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/13/25 22:26:40.976 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/13/25 22:43:23.135 (16m42.135s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-safe --namespace head-5073ae15-end-to-end-vm-restore-safe --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-safe\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:107 @ 10/13/25 22:43:23.135 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/13/25 22:43:23.137 (16m42.136s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 22:43:23.137 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/13/25 22:43:31.611 (8.474s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:115 @ 10/13/25 22:43:31.611 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:158 @ 10/13/25 22:43:31.611 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:348 @ 10/13/25 22:43:31.611 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:374 @ 10/13/25 22:43:31.611 + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400000e9f0>: the container "virtualization-controller" was not found: virtualization-controller-65d97ff7b4-25gpr the container "virtualization-controller" was restarted: virtualization-controller-765bd4c78-2mhvk the container "virtualization-controller" was restarted: virtualization-controller-765bd4c78-hfgs2 { errs: [ <*errors.joinError | 0x1400000e9c0>{ errs: [ <*errors.joinError | 0x1400000e990>{ errs: [ <*errors.errorString | 0x1400057e450>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-65d97ff7b4-25gpr", }, ], }, <*errors.errorString | 0x1400057e490>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-765bd4c78-2mhvk", }, ], }, <*errors.errorString | 0x1400057e4f0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-765bd4c78-hfgs2", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/13/25 22:43:32.579 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 22:43:31.611 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 22:43:31.612 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 22:43:31.612 [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400000e9f0>: the container "virtualization-controller" was not found: virtualization-controller-65d97ff7b4-25gpr the container "virtualization-controller" was restarted: virtualization-controller-765bd4c78-2mhvk the container "virtualization-controller" was restarted: virtualization-controller-765bd4c78-hfgs2 { errs: [ <*errors.joinError | 0x1400000e9c0>{ errs: [ <*errors.joinError | 0x1400000e990>{ errs: [ <*errors.errorString | 0x1400057e450>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-65d97ff7b4-25gpr", }, ], }, <*errors.errorString | 0x1400057e490>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-765bd4c78-2mhvk", }, ], }, <*errors.errorString | 0x1400057e4f0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-765bd4c78-hfgs2", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/13/25 22:43:32.579 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/13/25 22:43:32.579 (967ms) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/13/25 22:43:32.579 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/13/25 22:43:47.609 (15.03s) + + + \ No newline at end of file diff --git a/artifacts/debug-hostpath/junit-debug.xml b/artifacts/debug-hostpath/junit-debug.xml new file mode 100644 index 0000000000..a876cc5cfc --- /dev/null +++ b/artifacts/debug-hostpath/junit-debug.xml @@ -0,0 +1,691 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/06/25 20:06:10.139 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/06/25 20:06:13.887 (3.748s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/06/25 20:06:13.887 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/06/25 20:06:13.887 (0s) + + + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/06/25 20:06:13.888 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/06/25 20:06:15.858 (1.97s) > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/06/25 20:06:15.858 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/06/25 20:06:15.859 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/06/25 20:06:15.859 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/06/25 20:06:19.786 (3.927s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/06/25 20:06:19.786 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/06/25 20:06:19.786 (0s) + + + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/06/25 20:06:19.786 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/06/25 20:06:19.786 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/06/25 20:06:19.786 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:88 @ 10/06/25 20:06:19.786 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/06/25 20:06:26.044 (6.258s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/06/25 20:06:26.044 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/06/25 20:06:26.044 (0s) + + + [FAILED] Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-root-automatic-with-hotplug-standalone --namespace head-45bdb83d-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-root-automatic-with-hotplug-standalone\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/06/25 20:23:07.695 + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/06/25 20:06:26.044 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/06/25 20:06:26.044 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/06/25 20:06:26.044 STEP: VDs with consumers should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:99 @ 10/06/25 20:06:26.044 [FAILED] Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-root-automatic-with-hotplug-standalone --namespace head-45bdb83d-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-root-automatic-with-hotplug-standalone\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/06/25 20:23:07.695 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/06/25 20:23:07.695 (16m41.659s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/06/25 20:23:07.695 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/06/25 20:23:12.658 (4.964s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/06/25 20:23:12.659 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/06/25 20:23:12.659 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/06/25 20:23:12.659 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/06/25 20:23:12.659 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/06/25 20:23:12.659 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/06/25 20:23:12.659 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/06/25 20:23:12.659 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/06/25 20:23:12.659 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/06/25 20:23:12.659 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/06/25 20:23:12.659 + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:66 @ 10/06/25 20:23:14.645 + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:54 @ 10/06/25 20:23:12.659 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:66 @ 10/06/25 20:23:14.645 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:54 @ 10/06/25 20:23:14.645 (1.986s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/06/25 20:23:14.645 < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/06/25 20:23:19.159 (4.514s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:95 @ 10/06/25 20:23:19.159 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:106 @ 10/06/25 20:23:19.159 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:117 @ 10/06/25 20:23:19.159 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:128 @ 10/06/25 20:23:19.159 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:139 @ 10/06/25 20:23:19.159 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:159 @ 10/06/25 20:23:19.159 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:180 @ 10/06/25 20:23:19.159 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:224 @ 10/06/25 20:23:19.159 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:285 @ 10/06/25 20:23:19.159 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:331 @ 10/06/25 20:23:19.159 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:357 @ 10/06/25 20:23:19.159 + + + > Enter [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/06/25 20:23:19.159 < Exit [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/06/25 20:23:21.128 (1.968s) > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/06/25 20:23:21.128 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/06/25 20:23:21.128 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/06/25 20:23:21.128 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/06/25 20:23:27.966 (6.838s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/06/25 20:23:27.966 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/06/25 20:23:27.966 (0s) + + + [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/06/25 20:36:05.075 This is the Progress Report generated when the suite timeout occurred: VirtualMachineRestoreForce When the virtualization resources are applied checks the resources phase (Spec Runtime: 12m37.118s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 In [It] (Node Runtime: 12m37.118s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 At [By Step] `VirtualMachine` agent should be ready (Step Runtime: 12m37.118s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 Spec Goroutine goroutine 275 [sync.WaitGroup.Wait, 13 minutes] sync.runtime_SemacquireWaitGroup(0x1029c4b30?, 0x40?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x1400089e670) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000a84620, 0x2, 0x104434e70?}, {0x1041530d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x140005d37d0, {0x14000a27800, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1041530d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x140005d37d0, {0x14000a27800, 0x29}, {0x14000651440, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1041530d7, 0x2a}, {0x10411037b, 0x7}, {{0x0, 0x0, 0x0}, 0x140005d37d0, {0x14000a27800, 0x29}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x140005d37d0, {0x14000a27800, 0x29}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func21.4.2.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 | It("checks the resources phase", func() { | By("`VirtualMachine` agent should be ready", func() { > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: namespace, github.com/onsi/ginkgo/v2/internal.(*Suite).By(0x1400016f508, {0x10414a434, 0x26}, {0x14000ba3f48, 0x1, 0x1040a3584?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:323 github.com/onsi/ginkgo/v2.By({0x10414a434?, 0x104acc210?}, {0x1400048d748?, 0x140001a2690?, 0x1400048d758?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:600 > github.com/deckhouse/virtualization/tests/e2e.init.func21.4.2() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 | | It("checks the resources phase", func() { > By("`VirtualMachine` agent should be ready", func() { | WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 259 [syscall, 13 minutes] syscall.syscall6(0x1009f86b0?, 0x130262d00?, 0x106d1ca78?, 0x90?, 0x14000101808?, 0x140008bc5a0?, 0x14000982a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000982a98?, 0x102a466fc?, 0x90?, 0x104a34660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000274cb0?, 0x14000982ad4, 0x140009f86b0?, 0x14000274c40?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140000e7bc0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140004f3888?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140004ea300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140004ea300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002c74b0?, 0x104acc1e8?, 0x14000274b60?}}, {0x104aec808?, 0x14000274b60?}, {0x140009f41a0?, 0xc2311da345c668a8?}, {0x104ac5120, 0x140000e7a80}, {0x104ac5120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002c74b0?, 0x105fd3668?, 0x105fa9ee0?}}, {0x104aec808, 0x14000274b60}, {0x140009f41a0, 0xd0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x104af0e48?, 0x14000039740?}, {0x104110924?, 0x0?}}, {0x1041530d7, 0x2a}, {0x140009a6d20, 0x28}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 275 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 131 [sync.Cond.Wait, 3 minutes] sync.runtime_notifyListWait(0x140008c2348, 0xa9) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140008c2338) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140008c2330, {0x14000915000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x140008e20f0?}, {0x14000915000?, 0x1045a1640?, 0x14000188f20?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140005280c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140005280c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x140008f8008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x140008f8008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x140008f8008, {0x14000ac17e6, 0xda81a, 0x102b09c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x14000150848, {0x14000ac17e6, 0xda81a, 0xda81a}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000ac17e6?, 0x140008e4778?, 0x14000ac1694?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x1400091bda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140008e3110, {0x1302f1000, 0x14000610840}, {0x104ac5920, 0x14000119620}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 31 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 260 [syscall, 13 minutes] syscall.syscall6(0x10089e6b0?, 0x106d2bd70?, 0x106d1cf30?, 0x90?, 0x105fac3c0?, 0x140007fec60?, 0x140000b4a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140000b4a98?, 0x102a466fc?, 0x90?, 0x104a34660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002f6d20?, 0x140000b4ad4, 0x1400089e6b0?, 0x140002f6cb0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000521800) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000522008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140008c2600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140008c2600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002c74b0?, 0x104acc1e8?, 0x140002f6b60?}}, {0x104aec808?, 0x140002f6b60?}, {0x140008684e0?, 0xc2311da345c668a8?}, {0x104ac5120, 0x14000521700}, {0x104ac5120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002c74b0?, 0x105fd3668?, 0x105fa9ee0?}}, {0x104aec808, 0x140002f6b60}, {0x140008684e0, 0xc6}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x104af0e48?, 0x14000039740?}, {0x104110924?, 0x14000851e98?}}, {0x1041530d7, 0x2a}, {0x140009a6d49, 0x1e}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 275 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) + > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/06/25 20:23:27.966 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/06/25 20:23:27.966 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/06/25 20:23:27.966 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/06/25 20:23:27.966 [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/06/25 20:36:05.075 This is the Progress Report generated when the suite timeout occurred: VirtualMachineRestoreForce When the virtualization resources are applied checks the resources phase (Spec Runtime: 12m37.118s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 In [It] (Node Runtime: 12m37.118s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 At [By Step] `VirtualMachine` agent should be ready (Step Runtime: 12m37.118s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 Spec Goroutine goroutine 275 [sync.WaitGroup.Wait, 13 minutes] sync.runtime_SemacquireWaitGroup(0x1029c4b30?, 0x40?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x1400089e670) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000a84620, 0x2, 0x104434e70?}, {0x1041530d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x140005d37d0, {0x14000a27800, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1041530d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x140005d37d0, {0x14000a27800, 0x29}, {0x14000651440, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1041530d7, 0x2a}, {0x10411037b, 0x7}, {{0x0, 0x0, 0x0}, 0x140005d37d0, {0x14000a27800, 0x29}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x140005d37d0, {0x14000a27800, 0x29}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func21.4.2.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 | It("checks the resources phase", func() { | By("`VirtualMachine` agent should be ready", func() { > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: namespace, github.com/onsi/ginkgo/v2/internal.(*Suite).By(0x1400016f508, {0x10414a434, 0x26}, {0x14000ba3f48, 0x1, 0x1040a3584?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:323 github.com/onsi/ginkgo/v2.By({0x10414a434?, 0x104acc210?}, {0x1400048d748?, 0x140001a2690?, 0x1400048d758?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:600 > github.com/deckhouse/virtualization/tests/e2e.init.func21.4.2() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 | | It("checks the resources phase", func() { > By("`VirtualMachine` agent should be ready", func() { | WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 259 [syscall, 13 minutes] syscall.syscall6(0x1009f86b0?, 0x130262d00?, 0x106d1ca78?, 0x90?, 0x14000101808?, 0x140008bc5a0?, 0x14000982a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000982a98?, 0x102a466fc?, 0x90?, 0x104a34660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000274cb0?, 0x14000982ad4, 0x140009f86b0?, 0x14000274c40?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140000e7bc0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140004f3888?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140004ea300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140004ea300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002c74b0?, 0x104acc1e8?, 0x14000274b60?}}, {0x104aec808?, 0x14000274b60?}, {0x140009f41a0?, 0xc2311da345c668a8?}, {0x104ac5120, 0x140000e7a80}, {0x104ac5120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002c74b0?, 0x105fd3668?, 0x105fa9ee0?}}, {0x104aec808, 0x14000274b60}, {0x140009f41a0, 0xd0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x104af0e48?, 0x14000039740?}, {0x104110924?, 0x0?}}, {0x1041530d7, 0x2a}, {0x140009a6d20, 0x28}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 275 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 131 [sync.Cond.Wait, 3 minutes] sync.runtime_notifyListWait(0x140008c2348, 0xa9) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140008c2338) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140008c2330, {0x14000915000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x140008e20f0?}, {0x14000915000?, 0x1045a1640?, 0x14000188f20?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140005280c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140005280c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x140008f8008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x140008f8008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x140008f8008, {0x14000ac17e6, 0xda81a, 0x102b09c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x14000150848, {0x14000ac17e6, 0xda81a, 0xda81a}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000ac17e6?, 0x140008e4778?, 0x14000ac1694?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x1400091bda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140008e3110, {0x1302f1000, 0x14000610840}, {0x104ac5920, 0x14000119620}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 31 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 260 [syscall, 13 minutes] syscall.syscall6(0x10089e6b0?, 0x106d2bd70?, 0x106d1cf30?, 0x90?, 0x105fac3c0?, 0x140007fec60?, 0x140000b4a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140000b4a98?, 0x102a466fc?, 0x90?, 0x104a34660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002f6d20?, 0x140000b4ad4, 0x1400089e6b0?, 0x140002f6cb0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000521800) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000522008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140008c2600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140008c2600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002c74b0?, 0x104acc1e8?, 0x140002f6b60?}}, {0x104aec808?, 0x140002f6b60?}, {0x140008684e0?, 0xc2311da345c668a8?}, {0x104ac5120, 0x14000521700}, {0x104ac5120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002c74b0?, 0x105fd3668?, 0x105fa9ee0?}}, {0x104aec808, 0x140002f6b60}, {0x140008684e0, 0xc6}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x104af0e48?, 0x14000039740?}, {0x104110924?, 0x14000851e98?}}, {0x1041530d7, 0x2a}, {0x140009a6d49, 0x1e}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 275 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/06/25 20:36:05.08 (12m37.123s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/06/25 20:36:05.08 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/06/25 20:36:10.205 (5.125s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/06/25 20:36:10.207 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/06/25 20:36:10.207 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/06/25 20:36:10.207 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/06/25 20:36:10.393 (186ms) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/06/25 20:36:10.393 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/06/25 20:36:14.321 (3.928s) + + + \ No newline at end of file diff --git a/artifacts/debug-nfs/junit-debug.xml b/artifacts/debug-nfs/junit-debug.xml new file mode 100644 index 0000000000..8d5e5f95d1 --- /dev/null +++ b/artifacts/debug-nfs/junit-debug.xml @@ -0,0 +1,705 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/07/25 08:44:06.103 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/07/25 08:44:11.127 (5.025s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/07/25 08:44:11.127 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/07/25 08:44:11.127 (0s) + + + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:51 @ 10/07/25 08:44:11.13 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:51 @ 10/07/25 08:44:13.122 (1.991s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:61 @ 10/07/25 08:44:13.122 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:61 @ 10/07/25 08:44:41.21 (28.088s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/07/25 08:44:41.21 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/07/25 08:44:41.21 (0s) + + + > Enter [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:82 @ 10/07/25 08:44:41.21 < Exit [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:82 @ 10/07/25 08:44:42.333 (1.122s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/07/25 08:44:42.333 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/07/25 08:44:42.333 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:97 @ 10/07/25 08:44:42.333 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:98 @ 10/07/25 08:44:42.333 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:97 @ 10/07/25 08:44:43.725 (1.392s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/07/25 08:44:43.725 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/07/25 08:44:43.725 (0s) + + + > Enter [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:108 @ 10/07/25 08:44:43.725 STEP: CVIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:109 @ 10/07/25 08:44:43.725 < Exit [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:108 @ 10/07/25 08:44:44.882 (1.157s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/07/25 08:44:44.882 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/07/25 08:44:44.882 (0s) + + + > Enter [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:119 @ 10/07/25 08:44:44.882 STEP: VMClasses should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:120 @ 10/07/25 08:44:44.882 < Exit [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:119 @ 10/07/25 08:44:46.195 (1.313s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/07/25 08:44:46.195 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/07/25 08:44:46.195 (0s) + + + [FAILED] Timed out after 343.863s. Expected success, but got an error: <*errors.errorString | 0x14000666030>: cannot patch VMIP "head-45bdb83d-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-45bdb83d-vm-custom-ip { s: "cannot patch VMIP \"head-45bdb83d-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-45bdb83d-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/07/25 08:50:30.01 + > Enter [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 @ 10/07/25 08:44:46.195 [FAILED] Timed out after 343.863s. Expected success, but got an error: <*errors.errorString | 0x14000666030>: cannot patch VMIP "head-45bdb83d-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-45bdb83d-vm-custom-ip { s: "cannot patch VMIP \"head-45bdb83d-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-45bdb83d-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/07/25 08:50:30.01 < Exit [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 @ 10/07/25 08:50:30.01 (5m43.864s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/07/25 08:50:30.01 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/07/25 08:50:34.664 (4.654s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:137 @ 10/07/25 08:50:34.665 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:148 @ 10/07/25 08:50:34.665 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:158 @ 10/07/25 08:50:34.665 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:169 @ 10/07/25 08:50:34.665 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:180 @ 10/07/25 08:50:34.665 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:192 @ 10/07/25 08:50:34.665 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:216 @ 10/07/25 08:50:34.665 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:240 @ 10/07/25 08:50:34.665 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:263 @ 10/07/25 08:50:34.665 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:275 @ 10/07/25 08:50:34.665 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:293 @ 10/07/25 08:50:34.665 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:310 @ 10/07/25 08:50:34.665 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:334 @ 10/07/25 08:50:34.665 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:347 @ 10/07/25 08:50:34.665 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:351 @ 10/07/25 08:50:34.665 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:368 @ 10/07/25 08:50:34.665 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:381 @ 10/07/25 08:50:34.666 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:398 @ 10/07/25 08:50:34.666 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:437 @ 10/07/25 08:50:34.666 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:472 @ 10/07/25 08:50:34.666 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:490 @ 10/07/25 08:50:34.666 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:505 @ 10/07/25 08:50:34.666 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:521 @ 10/07/25 08:50:34.666 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:538 @ 10/07/25 08:50:34.666 + + + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/07/25 08:50:34.666 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/07/25 08:50:36.629 (1.963s) > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/07/25 08:50:36.629 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/07/25 08:50:36.629 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/07/25 08:50:36.629 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/07/25 08:50:40.253 (3.624s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/07/25 08:50:40.253 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/07/25 08:50:40.253 (0s) + + + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/07/25 08:50:40.253 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/07/25 08:50:40.253 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/07/25 08:50:40.253 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:88 @ 10/07/25 08:50:40.253 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/07/25 08:50:44.778 (4.525s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/07/25 08:50:44.778 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/07/25 08:50:44.778 (0s) + + + [FAILED] Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-root-automatic-with-hotplug-standalone --namespace head-45bdb83d-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-root-automatic-with-hotplug-standalone\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/07/25 09:07:26.926 + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/07/25 08:50:44.778 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/07/25 08:50:44.778 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/07/25 08:50:44.778 STEP: VDs with consumers should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:99 @ 10/07/25 08:50:44.778 [FAILED] Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-root-automatic-with-hotplug-standalone --namespace head-45bdb83d-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-root-automatic-with-hotplug-standalone\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/07/25 09:07:26.926 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/07/25 09:07:26.926 (16m42.131s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/07/25 09:07:26.926 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/07/25 09:07:31.835 (4.909s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/07/25 09:07:31.835 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/07/25 09:07:31.835 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/07/25 09:07:31.835 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/07/25 09:07:31.835 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/07/25 09:07:31.835 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/07/25 09:07:31.835 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/07/25 09:07:31.836 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/07/25 09:07:31.836 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/07/25 09:07:31.836 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/07/25 09:07:31.836 + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/07/25 09:07:31.836 < Exit [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/07/25 09:07:33.825 (1.989s) > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/07/25 09:07:33.825 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/07/25 09:07:33.825 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/07/25 09:07:33.825 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/07/25 09:07:38.434 (4.609s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/07/25 09:07:38.434 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/07/25 09:07:38.434 (0s) + + + [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/07/25 10:28:03.289 This is the Progress Report generated when the suite timeout occurred: VirtualMachineRestoreSafe When the virtualization resources are applied checks the resources phase (Spec Runtime: 6m20.211s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 In [It] (Node Runtime: 6m20.211s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 At [By Step] `VirtualMachine` agent should be ready (Step Runtime: 6m20.211s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 Spec Goroutine goroutine 330 [sync.WaitGroup.Wait, 7 minutes] sync.runtime_SemacquireWaitGroup(0x104ad8b30?, 0x30?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000c40750) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140006a58b0, 0x1, 0x106548e70?}, {0x1062670d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x140004ee240, {0x14000e23830, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1062670d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x140004ee240, {0x14000e23830, 0x28}, {0x14000efc9c0, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1062670d7, 0x2a}, {0x10622437b, 0x7}, {{0x0, 0x0, 0x0}, 0x140004ee240, {0x14000e23830, 0x28}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x140004ee240, {0x14000e23830, 0x28}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func22.4.2.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:107 | It("checks the resources phase", func() { | By("`VirtualMachine` agent should be ready", func() { > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: namespace, github.com/onsi/ginkgo/v2/internal.(*Suite).By(0x14000427508, {0x10625e434, 0x26}, {0x1400093ff50, 0x1, 0x14000614080?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:323 github.com/onsi/ginkgo/v2.By({0x10625e434?, 0x104c17b8c?}, {0x1400012cf50?, 0x1400012cf90?, 0x1400048e201?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:600 > github.com/deckhouse/virtualization/tests/e2e.init.func22.4.2() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 | | It("checks the resources phase", func() { > By("`VirtualMachine` agent should be ready", func() { | WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x1061c8050?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 7 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 340 [syscall, 7 minutes] syscall.syscall6(0x100c40790?, 0x108e3cb70?, 0x108e30a78?, 0x90?, 0x14000600008?, 0x140000647e0?, 0x140012faa68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140012faa98?, 0x104b5a6fc?, 0x90?, 0x106b48660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002c3570?, 0x140012faad4, 0x14000c40790?, 0x140002c3490?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140002a6fc0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x1400014f888?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x1400049e480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x1400049e480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002bd9a0?, 0x106be01e8?, 0x140002c32d0?}}, {0x106c00808?, 0x140002c32d0?}, {0x140008e8000?, 0xc2314a69e0304ba0?}, {0x106bd9120, 0x140002a6f40}, {0x106bd9120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002bd9a0?, 0x1080e7668?, 0x1080bdee0?}}, {0x106c00808, 0x140002c32d0}, {0x140008e8000, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106c04e48?, 0x14000722d50?}, {0x106224924?, 0x0?}}, {0x1062670d7, 0x2a}, {0x140002be520, 0x1d}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 330 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) + > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/07/25 09:07:38.434 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/07/25 09:07:38.434 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/07/25 09:07:38.434 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/07/25 09:07:38.434 [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/07/25 10:28:03.289 This is the Progress Report generated when the suite timeout occurred: VirtualMachineRestoreSafe When the virtualization resources are applied checks the resources phase (Spec Runtime: 6m20.211s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 In [It] (Node Runtime: 6m20.211s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 At [By Step] `VirtualMachine` agent should be ready (Step Runtime: 6m20.211s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 Spec Goroutine goroutine 330 [sync.WaitGroup.Wait, 7 minutes] sync.runtime_SemacquireWaitGroup(0x104ad8b30?, 0x30?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000c40750) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140006a58b0, 0x1, 0x106548e70?}, {0x1062670d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x140004ee240, {0x14000e23830, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1062670d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x140004ee240, {0x14000e23830, 0x28}, {0x14000efc9c0, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1062670d7, 0x2a}, {0x10622437b, 0x7}, {{0x0, 0x0, 0x0}, 0x140004ee240, {0x14000e23830, 0x28}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x140004ee240, {0x14000e23830, 0x28}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func22.4.2.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:107 | It("checks the resources phase", func() { | By("`VirtualMachine` agent should be ready", func() { > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: namespace, github.com/onsi/ginkgo/v2/internal.(*Suite).By(0x14000427508, {0x10625e434, 0x26}, {0x1400093ff50, 0x1, 0x14000614080?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:323 github.com/onsi/ginkgo/v2.By({0x10625e434?, 0x104c17b8c?}, {0x1400012cf50?, 0x1400012cf90?, 0x1400048e201?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:600 > github.com/deckhouse/virtualization/tests/e2e.init.func22.4.2() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 | | It("checks the resources phase", func() { > By("`VirtualMachine` agent should be ready", func() { | WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x1061c8050?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 7 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 340 [syscall, 7 minutes] syscall.syscall6(0x100c40790?, 0x108e3cb70?, 0x108e30a78?, 0x90?, 0x14000600008?, 0x140000647e0?, 0x140012faa68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140012faa98?, 0x104b5a6fc?, 0x90?, 0x106b48660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002c3570?, 0x140012faad4, 0x14000c40790?, 0x140002c3490?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140002a6fc0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x1400014f888?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x1400049e480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x1400049e480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002bd9a0?, 0x106be01e8?, 0x140002c32d0?}}, {0x106c00808?, 0x140002c32d0?}, {0x140008e8000?, 0xc2314a69e0304ba0?}, {0x106bd9120, 0x140002a6f40}, {0x106bd9120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002bd9a0?, 0x1080e7668?, 0x1080bdee0?}}, {0x106c00808, 0x140002c32d0}, {0x140008e8000, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106c04e48?, 0x14000722d50?}, {0x106224924?, 0x0?}}, {0x1062670d7, 0x2a}, {0x140002be520, 0x1d}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 330 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/07/25 10:28:03.294 (6m20.216s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/07/25 10:28:03.294 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/07/25 10:28:07.737 (4.445s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x14000980318>: the container "virtualization-controller" was not found: virtualization-controller-65dd99bb48-bdflv { errs: [ <*errors.errorString | 0x140002c6440>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-65dd99bb48-bdflv", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/07/25 10:28:08.584 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/07/25 10:28:07.74 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/07/25 10:28:07.74 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/07/25 10:28:07.74 [FAILED] Expected success, but got an error: <*errors.joinError | 0x14000980318>: the container "virtualization-controller" was not found: virtualization-controller-65dd99bb48-bdflv { errs: [ <*errors.errorString | 0x140002c6440>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-65dd99bb48-bdflv", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/07/25 10:28:08.584 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/07/25 10:28:08.584 (844ms) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/07/25 10:28:08.584 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/07/25 10:28:17.502 (8.92s) + + + \ No newline at end of file diff --git a/artifacts/fixed-cephrbd-20251011-211703-16149/junit.xml b/artifacts/fixed-cephrbd-20251011-211703-16149/junit.xml new file mode 100644 index 0000000000..d0872c6537 --- /dev/null +++ b/artifacts/fixed-cephrbd-20251011-211703-16149/junit.xml @@ -0,0 +1,672 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/11/25 22:26:28.228 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/11/25 22:26:32.805 (4.577s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/11/25 22:26:32.805 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/11/25 22:26:32.805 (1ms) + + + > Enter [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/11/25 22:26:32.808 < Exit [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/11/25 22:26:36.361 (3.554s) > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/11/25 22:26:36.361 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/11/25 22:26:36.361 (0s) > Enter [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/11/25 22:26:36.361 STEP: Create a vmip automatically and check its binding with a lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:81 @ 10/11/25 22:26:36.361 STEP: Remove label from the lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:87 @ 10/11/25 22:26:36.732 STEP: Wait for the label to be restored by the controller - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:96 @ 10/11/25 22:26:36.858 < Exit [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/11/25 22:26:37.097 (735ms) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/11/25 22:26:37.097 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/11/25 22:26:37.097 (0s) + + + > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/11/25 22:26:37.097 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/11/25 22:26:37.097 (0s) > Enter [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/11/25 22:26:37.097 STEP: Create an intermediate vmip automatically to allocate a new ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:109 @ 10/11/25 22:26:37.097 STEP: Delete the intermediate vmip automatically and check that the lease is released - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:115 @ 10/11/25 22:26:37.462 STEP: Reuse the released lease with a static vmip - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:124 @ 10/11/25 22:26:37.705 STEP: Delete the static vmip and lease, then create another static vmip with this ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:132 @ 10/11/25 22:26:38.078 < Exit [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/11/25 22:26:38.711 (1.614s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/11/25 22:26:38.711 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/11/25 22:26:38.711 (0s) + + + > Enter [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/11/25 22:26:38.712 < Exit [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/11/25 22:26:40.674 (1.963s) > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 22:26:40.674 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 22:26:40.674 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/11/25 22:26:40.674 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/11/25 22:26:43.293 (2.619s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 22:26:43.293 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 22:26:43.293 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 22:26:43.294 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 22:26:43.294 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/11/25 22:26:43.294 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:68 @ 10/11/25 22:26:43.294 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/11/25 22:27:00.441 (17.148s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 22:27:00.441 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 22:27:00.441 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 22:27:00.441 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 22:27:00.441 (0s) > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/11/25 22:27:00.441 STEP: VM should be in Running phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:79 @ 10/11/25 22:27:00.441 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/11/25 22:27:03.328 (2.886s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 22:27:03.328 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 22:27:03.328 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 22:27:03.328 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 22:27:03.328 (0s) > Enter [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/11/25 22:27:03.328 < Exit [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/11/25 22:27:03.328 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 22:27:03.328 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 22:27:03.328 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 22:27:03.328 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 22:27:03.328 (0s) > Enter [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/11/25 22:27:03.328 < Exit [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/11/25 22:27:03.328 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 22:27:03.328 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 22:27:03.328 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 22:27:03.329 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 22:27:03.329 (0s) > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/11/25 22:27:03.329 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/11/25 22:27:03.329 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/11/25 22:28:11.732 (1m8.405s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/11/25 22:28:11.733 (1m8.405s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 22:28:11.733 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 22:28:11.733 (0s) + + + > Enter [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/11/25 22:28:11.733 < Exit [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/11/25 22:28:11.734 (1ms) > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 22:28:11.734 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 22:28:11.734 (0s) > Enter [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/11/25 22:28:11.734 < Exit [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/11/25 22:28:12.969 (1.235s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 22:28:12.97 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 22:28:12.97 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 22:28:12.97 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 22:28:12.97 (0s) > Enter [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/11/25 22:28:12.97 STEP: Project should be deployed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:73 @ 10/11/25 22:28:12.97 < Exit [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/11/25 22:28:15.728 (2.758s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 22:28:15.728 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 22:28:15.728 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 22:28:15.728 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 22:28:15.728 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/11/25 22:28:15.728 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/11/25 22:28:18.73 (3.002s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 22:28:18.73 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 22:28:18.73 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 22:28:18.731 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 22:28:18.731 (0s) > Enter [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/11/25 22:28:18.731 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/11/25 22:28:18.731 < Exit [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/11/25 22:28:25.804 (7.074s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 22:28:25.804 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 22:28:25.804 (0s) + + + [INTERRUPTED] Interrupted by User In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/11/25 22:28:31.47 This is the Progress Report generated when the interrupt was received: ImporterNetworkPolicy When resources are applied When virtual disks are applied (Spec Runtime: 5.665s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 In [It] (Node Runtime: 5.665s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 At [By Step] VDs should be in Ready phases (Step Runtime: 5.665s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 Spec Goroutine goroutine 234 [sync.WaitGroup.Wait] sync.runtime_SemacquireWaitGroup(0x104cb8b30?, 0x10?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000a18da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14001082520, 0x2, 0x106731e58?}, {0x10644a204, 0x28}, {{0x0, 0x0, 0x0}, 0x140005fea80, {0x14000057a10, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x10644a204, 0x28}, {{0x0, 0x0, 0x0}, 0x140005fea80, {0x14000057a10, 0x30}, {0x14000055900, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x10644a204, 0x28}, {0x106408eef, 0x5}, {{0x0, 0x0, 0x0}, 0x140005fea80, {0x14000057a10, 0x30}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func7.7({0x10640795c?, 0x0?}, {0x10644a204, 0x28}, {0x106408eef, 0x5}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 | func(resourceShortName string, resource kc.Resource, phase string) { | By(fmt.Sprintf("%ss should be in %s phases", resourceShortName, phase)) > WaitPhaseByLabel(resource, phase, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, reflect.Value.call({0x1069ae7e0?, 0x14000404e70?, 0x140000b4e18?}, {0x106408530, 0x4}, {0x140002c81e0, 0x3, 0x1050e11b0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/reflect/value.go:581 reflect.Value.Call({0x1069ae7e0?, 0x14000404e70?, 0x106e18ca0?}, {0x140002c81e0?, 0x0?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/reflect/value.go:365 > github.com/onsi/ginkgo/v2.invokeFunction({0x1069ae7e0, 0x14000404e70}, {0x1400061e800, 0x3, 0x14000771000?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/table_dsl.go:339 | } | > return reflect.ValueOf(function).Call(inValues) | } | > github.com/onsi/ginkgo/v2.generateTable.func2.2() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/table_dsl.go:303 | panic(err) | } > invokeFunction(internalBody, entry.parameters) | }) | } > github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x14000653c50?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 | | body := arg.(func()) > return func(SpecContext) { body() }, false | } | > github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 | }() | > node.Body(sc) | finished = true | }() > github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 | failureC := make(chan types.Failure) | > go func() { | finished := false | defer func() { Goroutines of Interest goroutine 28 [select, 2 minutes] > github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal.func1() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:32 | go func() { | for { > select { | case <-signalChannel: | handler() > github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:30 | } | ctx, cancel := context.WithCancel(context.Background()) > go func() { | for { | select { goroutine 100 [sync.Cond.Wait] sync.runtime_notifyListWait(0x14000764648, 0x1c) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x14000764638) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x14000764630, {0x1400099c000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400099a150?}, {0x1400099c000?, 0x1068a1b40?, 0x140010b9080?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140000a2ae0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140000a2ae0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x140007d6008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x140007d6008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x140007d6008, {0x14000b060db, 0xf9f25, 0x104dfdc74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x14000329348, {0x14000b060db, 0xf9f25, 0xf9f25}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000b060db?, 0x1400063d7c8?, 0x14000b05e61?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x140009c9da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140004323c0, {0x114708798, 0x1400099a030}, {0x106dc75a0, 0x140000753e0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 71 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 39 [select] > github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts.func2(0x0?) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:131 | var interruptCause InterruptCause | for { > select { | case <-signalChannel: | interruptCause = InterruptCauseSignal > github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:128 | } | > go func(abortChannel chan interface{}) { | var interruptCause InterruptCause | for { goroutine 102 [sync.Cond.Wait] sync.runtime_notifyListWait(0x140000dcf48, 0xa0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140000dcf38) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140000dcf30, {0x140005c4000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400099a150?}, {0x140005c4000?, 0x1068a1b40?, 0x1400070d550?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x14000075f20) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x14000075f20) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000aa0008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000aa0008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000aa0008, {0x14000c19a84, 0xe657c, 0x104dfdc74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400026c008, {0x14000c19a84, 0xe657c, 0xe657c}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000c19a84?, 0x140004ebdf8?, 0x14000c19963?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x14000033da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140004323c0, {0x114708798, 0x1400099af60}, {0x106dc75a0, 0x140000753e0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 71 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 239 [syscall] syscall.syscall6(0x100a18de0?, 0x11465e308?, 0x109028a78?, 0x90?, 0x140004a1808?, 0x1400028e2d0?, 0x14000982a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000982a88?, 0x104d3a6fc?, 0x90?, 0x106d36120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400019d110?, 0x14000982ac4, 0x14000a18de0?, 0x1400019d0a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000754640) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140000dbc08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140000dc480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140000dc480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a3490?, 0x106dcde88?, 0x1400019cfc0?}}, {0x106dee488?, 0x1400019cfc0?}, {0x14000564240?, 0xc232cad5b38f99a8?}, {0x106dc6d80, 0x140007545c0}, {0x106dc6d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a3490?, 0x1082db728?, 0x1082b1fa0?}}, {0x106dee488, 0x1400019cfc0}, {0x14000564240, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106df2a38?, 0x140004bd350?}, {0x10640bfba?, 0x25?}}, {0x10644a204, 0x28}, {0x14000a05901, 0x14}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 234 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 238 [syscall] syscall.syscall6(0x1009147b0?, 0x11479d048?, 0x109028108?, 0x90?, 0x140000a5008?, 0x140010d21b0?, 0x140000b4a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140000b4a88?, 0x104d3a6fc?, 0x90?, 0x106d36120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029c540?, 0x140000b4ac4, 0x140009147b0?, 0x1400029c4d0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x1400059c200) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140010d5408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000628180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000628180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a3490?, 0x106dcde88?, 0x1400029c3f0?}}, {0x106dee488?, 0x1400029c3f0?}, {0x140010fe000?, 0xc232cad5b38fd440?}, {0x106dc6d80, 0x1400059c180}, {0x106dc6d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a3490?, 0x1082db728?, 0x1082b1fa0?}}, {0x106dee488, 0x1400029c3f0}, {0x140010fe000, 0xbb}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106df2a38?, 0x140004bd350?}, {0x10640bfba?, 0x140000b4e88?}}, {0x10644a204, 0x28}, {0x14000a058f0, 0x10}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 234 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) Other Goroutines goroutine 8 [running] github.com/onsi/ginkgo/v2/internal.extractRunningGoroutines() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:181 github.com/onsi/ginkgo/v2/internal.NewProgressReport(_, {{0x14001082360, 0x2, 0x2}, {0x1400019ca10, 0x2, 0x2}, {0x14000a139e0, 0x2, 0x2}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:75 github.com/onsi/ginkgo/v2/internal.(*Suite).generateProgressReport(_, _) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:381 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode(_, {0x5e, 0x4, {0x106435596, 0x1e}, 0x140002a3560, {{0x1074f578d, 0x5d}, 0x66, {0x0, ...}, ...}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:980 github.com/onsi/ginkgo/v2/internal.(*group).attemptSpec(0x140008077f0, 0x1, {{0x1400051b008?, 0x1400019ca10?, 0x2?}, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:199 github.com/onsi/ginkgo/v2/internal.(*group).run(0x140008077f0, {0x140006e2000?, 0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:349 github.com/onsi/ginkgo/v2/internal.(*Suite).runSpecs(0x1400047f508, {0x106408f2b, 0x5}, {0x1082db728, 0x0, 0x0}, {0x1400005a0f4, 0x3d}, 0x0, {0x14000092c08, ...}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:489 github.com/onsi/ginkgo/v2/internal.(*Suite).Run(_, {_, _}, {_, _, _}, {_, _}, _, {_, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:130 github.com/onsi/ginkgo/v2.RunSpecs({0x106dc60a0, 0x14000502e00}, {0x106408f2b, 0x5}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:300 github.com/deckhouse/virtualization/tests/e2e.TestE2E(0x14000502e00) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:76 testing.tRunner(0x14000502e00, 0x106db5c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.(*T).Run in goroutine 1 /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1997 goroutine 1 [chan receive, 2 minutes] testing.(*T).Run(0x14000502c40, {0x10640b99a?, 0x14000597b38?}, 0x106db5c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2005 testing.runTests.func1(0x14000502c40) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2477 testing.tRunner(0x14000502c40, 0x14000597c68) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.runTests(0x140004bc7b0, {0x108260fd0, 0x1, 0x1}, {0x140004e5780?, 0xf?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2475 testing.(*M).Run(0x140005af0e0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2337 main.main() _testmain.go:45 goroutine 244 [IO wait] internal/poll.runtime_pollWait(0x1146fb600, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140000754a0?, 0x140009d8000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140000754a0, {0x140009d8000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140004281b8, {0x140009d8000?, 0x14001155da8?, 0x104c7ba18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x106dc6d80, 0x1400059c1c0}, {0x106dc5ec8, 0x140004980c8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x106dc6d80, 0x1400059c1c0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x108261230?, {0x106dc6d80?, 0x1400059c1c0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x106dc6d80, 0x1400059c1c0}, {0x106dc5fc0, 0x140004281b8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 238 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 53 [IO wait] internal/poll.runtime_pollWait(0x1146fba00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140005b0f80?, 0x1400083e500?, 0x0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140005b0f80, {0x1400083e500, 0x2500, 0x2500}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 net.(*netFD).Read(0x140005b0f80, {0x1400083e500?, 0x14000031818?, 0x104c7eb28?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/net/fd_posix.go:68 net.(*conn).Read(0x14000428000, {0x1400083e500?, 0x30083e500?, 0x1145aa8e0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/net/net.go:196 crypto/tls.(*atLeastReader).Read(0x140000de390, {0x1400083e500?, 0x140000318d8?, 0x104f93ed4?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:816 bytes.(*Buffer).ReadFrom(0x140002ad7a8, {0x106dc7de0, 0x140000de390}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bytes/buffer.go:217 crypto/tls.(*Conn).readFromUntil(0x140002ad508, {0x11482a200, 0x14000428000}, 0x14000031980?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:838 crypto/tls.(*Conn).readRecordOrCCS(0x140002ad508, 0x0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:627 crypto/tls.(*Conn).readRecord(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:589 crypto/tls.(*Conn).Read(0x140002ad508, {0x140006e7000, 0x1000, 0x1400013a480?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:1392 bufio.(*Reader).Read(0x140006e0060, {0x140002a44a0, 0x9, 0x1052e3f40?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:245 io.ReadAtLeast({0x106dc63c0, 0x140006e0060}, {0x140002a44a0, 0x9, 0x9}, 0x9) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:335 io.ReadFull(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:354 golang.org/x/net/http2.readFrameHeader({0x140002a44a0, 0x9, 0x14000031de8?}, {0x106dc63c0?, 0x140006e0060?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/frame.go:242 golang.org/x/net/http2.(*Framer).ReadFrame(0x140002a4460) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/frame.go:506 golang.org/x/net/http2.(*clientConnReadLoop).run(0x14000031f98) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2258 golang.org/x/net/http2.(*ClientConn).readLoop(0x14000102a80) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2127 golang.org/x/net/http2.(*Transport).newClientConn in goroutine 52 /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:912 goroutine 82 [syscall] os/signal.signal_recv() /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sigqueue.go:149 os/signal.loop() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal_unix.go:23 os/signal.Notify.func1.1 in goroutine 8 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal.go:152 goroutine 101 [select, 2 minutes] golang.org/x/net/http2.(*clientStream).writeRequest(0x140000dcf00, 0x14000573180, 0x0) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1570 golang.org/x/net/http2.(*clientStream).doRequest(0x140000dcf00, 0x104cedb64?, 0x14000628780?) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1431 golang.org/x/net/http2.(*ClientConn).roundTrip in goroutine 71 /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1336 goroutine 64 [select, 2 minutes] golang.org/x/net/http2.(*clientStream).writeRequest(0x14000764600, 0x1400087e640, 0x0) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1570 golang.org/x/net/http2.(*clientStream).doRequest(0x14000764600, 0x0?, 0x0?) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1431 golang.org/x/net/http2.(*ClientConn).roundTrip in goroutine 71 /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1336 goroutine 240 [IO wait] internal/poll.runtime_pollWait(0x1146fc000, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140007306c0?, 0x14000d10000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140007306c0, {0x14000d10000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000554a58, {0x14000d10000?, 0x14000983da8?, 0x104c7ba18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x106dc6d80, 0x140007545c0}, {0x106dc5ec8, 0x14000d0e000}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x106dc6d80, 0x140007545c0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x108261230?, {0x106dc6d80?, 0x140007545c0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x106dc6d80, 0x140007545c0}, {0x106dc5fc0, 0x14000554a58}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 239 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 241 [IO wait] internal/poll.runtime_pollWait(0x1146fb200, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000730780?, 0x14000a32000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000730780, {0x14000a32000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000554a70, {0x14000a32000?, 0x14000984da8?, 0x104c7ba18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x106dc6d80, 0x14000754600}, {0x106dc5ec8, 0x1400011c3f8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x106dc6d80, 0x14000754600}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x108261230?, {0x106dc6d80?, 0x14000754600?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x106dc6d80, 0x14000754600}, {0x106dc5fc0, 0x14000554a70}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 239 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 290 [select] os/exec.(*Cmd).watchCtx(0x140000dc480, 0x1400057ed20) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 239 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 243 [IO wait] internal/poll.runtime_pollWait(0x1146fbc00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000074d80?, 0x14000a2a000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000074d80, {0x14000a2a000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000428188, {0x14000a2a000?, 0x14001154da8?, 0x104c7ba18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x106dc6d80, 0x1400059c180}, {0x106dc5ec8, 0x1400011c3f0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x106dc6d80, 0x1400059c180}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x108261230?, {0x106dc6d80?, 0x1400059c180?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x106dc6d80, 0x1400059c180}, {0x106dc5fc0, 0x14000428188}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 238 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 245 [select] os/exec.(*Cmd).watchCtx(0x14000628180, 0x140006c6460) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 238 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 22:28:25.804 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 22:28:25.804 (0s) > Enter [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/11/25 22:28:25.804 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/11/25 22:28:25.804 [INTERRUPTED] Interrupted by User In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/11/25 22:28:31.47 This is the Progress Report generated when the interrupt was received: ImporterNetworkPolicy When resources are applied When virtual disks are applied (Spec Runtime: 5.665s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 In [It] (Node Runtime: 5.665s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 At [By Step] VDs should be in Ready phases (Step Runtime: 5.665s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 Spec Goroutine goroutine 234 [sync.WaitGroup.Wait] sync.runtime_SemacquireWaitGroup(0x104cb8b30?, 0x10?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000a18da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14001082520, 0x2, 0x106731e58?}, {0x10644a204, 0x28}, {{0x0, 0x0, 0x0}, 0x140005fea80, {0x14000057a10, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x10644a204, 0x28}, {{0x0, 0x0, 0x0}, 0x140005fea80, {0x14000057a10, 0x30}, {0x14000055900, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x10644a204, 0x28}, {0x106408eef, 0x5}, {{0x0, 0x0, 0x0}, 0x140005fea80, {0x14000057a10, 0x30}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func7.7({0x10640795c?, 0x0?}, {0x10644a204, 0x28}, {0x106408eef, 0x5}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 | func(resourceShortName string, resource kc.Resource, phase string) { | By(fmt.Sprintf("%ss should be in %s phases", resourceShortName, phase)) > WaitPhaseByLabel(resource, phase, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, reflect.Value.call({0x1069ae7e0?, 0x14000404e70?, 0x140000b4e18?}, {0x106408530, 0x4}, {0x140002c81e0, 0x3, 0x1050e11b0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/reflect/value.go:581 reflect.Value.Call({0x1069ae7e0?, 0x14000404e70?, 0x106e18ca0?}, {0x140002c81e0?, 0x0?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/reflect/value.go:365 > github.com/onsi/ginkgo/v2.invokeFunction({0x1069ae7e0, 0x14000404e70}, {0x1400061e800, 0x3, 0x14000771000?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/table_dsl.go:339 | } | > return reflect.ValueOf(function).Call(inValues) | } | > github.com/onsi/ginkgo/v2.generateTable.func2.2() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/table_dsl.go:303 | panic(err) | } > invokeFunction(internalBody, entry.parameters) | }) | } > github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x14000653c50?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 | | body := arg.(func()) > return func(SpecContext) { body() }, false | } | > github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 | }() | > node.Body(sc) | finished = true | }() > github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 | failureC := make(chan types.Failure) | > go func() { | finished := false | defer func() { Goroutines of Interest goroutine 28 [select, 2 minutes] > github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal.func1() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:32 | go func() { | for { > select { | case <-signalChannel: | handler() > github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:30 | } | ctx, cancel := context.WithCancel(context.Background()) > go func() { | for { | select { goroutine 100 [sync.Cond.Wait] sync.runtime_notifyListWait(0x14000764648, 0x1c) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x14000764638) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x14000764630, {0x1400099c000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400099a150?}, {0x1400099c000?, 0x1068a1b40?, 0x140010b9080?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140000a2ae0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140000a2ae0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x140007d6008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x140007d6008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x140007d6008, {0x14000b060db, 0xf9f25, 0x104dfdc74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x14000329348, {0x14000b060db, 0xf9f25, 0xf9f25}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000b060db?, 0x1400063d7c8?, 0x14000b05e61?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x140009c9da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140004323c0, {0x114708798, 0x1400099a030}, {0x106dc75a0, 0x140000753e0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 71 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 39 [select] > github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts.func2(0x0?) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:131 | var interruptCause InterruptCause | for { > select { | case <-signalChannel: | interruptCause = InterruptCauseSignal > github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:128 | } | > go func(abortChannel chan interface{}) { | var interruptCause InterruptCause | for { goroutine 102 [sync.Cond.Wait] sync.runtime_notifyListWait(0x140000dcf48, 0xa0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140000dcf38) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140000dcf30, {0x140005c4000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400099a150?}, {0x140005c4000?, 0x1068a1b40?, 0x1400070d550?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x14000075f20) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x14000075f20) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000aa0008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000aa0008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000aa0008, {0x14000c19a84, 0xe657c, 0x104dfdc74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400026c008, {0x14000c19a84, 0xe657c, 0xe657c}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000c19a84?, 0x140004ebdf8?, 0x14000c19963?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x14000033da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140004323c0, {0x114708798, 0x1400099af60}, {0x106dc75a0, 0x140000753e0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 71 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 239 [syscall] syscall.syscall6(0x100a18de0?, 0x11465e308?, 0x109028a78?, 0x90?, 0x140004a1808?, 0x1400028e2d0?, 0x14000982a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000982a88?, 0x104d3a6fc?, 0x90?, 0x106d36120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400019d110?, 0x14000982ac4, 0x14000a18de0?, 0x1400019d0a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000754640) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140000dbc08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140000dc480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140000dc480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a3490?, 0x106dcde88?, 0x1400019cfc0?}}, {0x106dee488?, 0x1400019cfc0?}, {0x14000564240?, 0xc232cad5b38f99a8?}, {0x106dc6d80, 0x140007545c0}, {0x106dc6d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a3490?, 0x1082db728?, 0x1082b1fa0?}}, {0x106dee488, 0x1400019cfc0}, {0x14000564240, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106df2a38?, 0x140004bd350?}, {0x10640bfba?, 0x25?}}, {0x10644a204, 0x28}, {0x14000a05901, 0x14}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 234 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 238 [syscall] syscall.syscall6(0x1009147b0?, 0x11479d048?, 0x109028108?, 0x90?, 0x140000a5008?, 0x140010d21b0?, 0x140000b4a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140000b4a88?, 0x104d3a6fc?, 0x90?, 0x106d36120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029c540?, 0x140000b4ac4, 0x140009147b0?, 0x1400029c4d0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x1400059c200) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140010d5408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000628180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000628180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a3490?, 0x106dcde88?, 0x1400029c3f0?}}, {0x106dee488?, 0x1400029c3f0?}, {0x140010fe000?, 0xc232cad5b38fd440?}, {0x106dc6d80, 0x1400059c180}, {0x106dc6d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a3490?, 0x1082db728?, 0x1082b1fa0?}}, {0x106dee488, 0x1400029c3f0}, {0x140010fe000, 0xbb}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106df2a38?, 0x140004bd350?}, {0x10640bfba?, 0x140000b4e88?}}, {0x10644a204, 0x28}, {0x14000a058f0, 0x10}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 234 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) Other Goroutines goroutine 8 [running] github.com/onsi/ginkgo/v2/internal.extractRunningGoroutines() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:181 github.com/onsi/ginkgo/v2/internal.NewProgressReport(_, {{0x14001082360, 0x2, 0x2}, {0x1400019ca10, 0x2, 0x2}, {0x14000a139e0, 0x2, 0x2}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:75 github.com/onsi/ginkgo/v2/internal.(*Suite).generateProgressReport(_, _) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:381 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode(_, {0x5e, 0x4, {0x106435596, 0x1e}, 0x140002a3560, {{0x1074f578d, 0x5d}, 0x66, {0x0, ...}, ...}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:980 github.com/onsi/ginkgo/v2/internal.(*group).attemptSpec(0x140008077f0, 0x1, {{0x1400051b008?, 0x1400019ca10?, 0x2?}, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:199 github.com/onsi/ginkgo/v2/internal.(*group).run(0x140008077f0, {0x140006e2000?, 0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:349 github.com/onsi/ginkgo/v2/internal.(*Suite).runSpecs(0x1400047f508, {0x106408f2b, 0x5}, {0x1082db728, 0x0, 0x0}, {0x1400005a0f4, 0x3d}, 0x0, {0x14000092c08, ...}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:489 github.com/onsi/ginkgo/v2/internal.(*Suite).Run(_, {_, _}, {_, _, _}, {_, _}, _, {_, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:130 github.com/onsi/ginkgo/v2.RunSpecs({0x106dc60a0, 0x14000502e00}, {0x106408f2b, 0x5}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:300 github.com/deckhouse/virtualization/tests/e2e.TestE2E(0x14000502e00) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:76 testing.tRunner(0x14000502e00, 0x106db5c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.(*T).Run in goroutine 1 /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1997 goroutine 1 [chan receive, 2 minutes] testing.(*T).Run(0x14000502c40, {0x10640b99a?, 0x14000597b38?}, 0x106db5c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2005 testing.runTests.func1(0x14000502c40) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2477 testing.tRunner(0x14000502c40, 0x14000597c68) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.runTests(0x140004bc7b0, {0x108260fd0, 0x1, 0x1}, {0x140004e5780?, 0xf?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2475 testing.(*M).Run(0x140005af0e0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2337 main.main() _testmain.go:45 goroutine 244 [IO wait] internal/poll.runtime_pollWait(0x1146fb600, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140000754a0?, 0x140009d8000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140000754a0, {0x140009d8000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140004281b8, {0x140009d8000?, 0x14001155da8?, 0x104c7ba18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x106dc6d80, 0x1400059c1c0}, {0x106dc5ec8, 0x140004980c8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x106dc6d80, 0x1400059c1c0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x108261230?, {0x106dc6d80?, 0x1400059c1c0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x106dc6d80, 0x1400059c1c0}, {0x106dc5fc0, 0x140004281b8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 238 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 53 [IO wait] internal/poll.runtime_pollWait(0x1146fba00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140005b0f80?, 0x1400083e500?, 0x0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140005b0f80, {0x1400083e500, 0x2500, 0x2500}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 net.(*netFD).Read(0x140005b0f80, {0x1400083e500?, 0x14000031818?, 0x104c7eb28?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/net/fd_posix.go:68 net.(*conn).Read(0x14000428000, {0x1400083e500?, 0x30083e500?, 0x1145aa8e0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/net/net.go:196 crypto/tls.(*atLeastReader).Read(0x140000de390, {0x1400083e500?, 0x140000318d8?, 0x104f93ed4?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:816 bytes.(*Buffer).ReadFrom(0x140002ad7a8, {0x106dc7de0, 0x140000de390}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bytes/buffer.go:217 crypto/tls.(*Conn).readFromUntil(0x140002ad508, {0x11482a200, 0x14000428000}, 0x14000031980?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:838 crypto/tls.(*Conn).readRecordOrCCS(0x140002ad508, 0x0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:627 crypto/tls.(*Conn).readRecord(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:589 crypto/tls.(*Conn).Read(0x140002ad508, {0x140006e7000, 0x1000, 0x1400013a480?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:1392 bufio.(*Reader).Read(0x140006e0060, {0x140002a44a0, 0x9, 0x1052e3f40?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:245 io.ReadAtLeast({0x106dc63c0, 0x140006e0060}, {0x140002a44a0, 0x9, 0x9}, 0x9) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:335 io.ReadFull(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:354 golang.org/x/net/http2.readFrameHeader({0x140002a44a0, 0x9, 0x14000031de8?}, {0x106dc63c0?, 0x140006e0060?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/frame.go:242 golang.org/x/net/http2.(*Framer).ReadFrame(0x140002a4460) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/frame.go:506 golang.org/x/net/http2.(*clientConnReadLoop).run(0x14000031f98) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2258 golang.org/x/net/http2.(*ClientConn).readLoop(0x14000102a80) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2127 golang.org/x/net/http2.(*Transport).newClientConn in goroutine 52 /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:912 goroutine 82 [syscall] os/signal.signal_recv() /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sigqueue.go:149 os/signal.loop() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal_unix.go:23 os/signal.Notify.func1.1 in goroutine 8 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal.go:152 goroutine 101 [select, 2 minutes] golang.org/x/net/http2.(*clientStream).writeRequest(0x140000dcf00, 0x14000573180, 0x0) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1570 golang.org/x/net/http2.(*clientStream).doRequest(0x140000dcf00, 0x104cedb64?, 0x14000628780?) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1431 golang.org/x/net/http2.(*ClientConn).roundTrip in goroutine 71 /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1336 goroutine 64 [select, 2 minutes] golang.org/x/net/http2.(*clientStream).writeRequest(0x14000764600, 0x1400087e640, 0x0) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1570 golang.org/x/net/http2.(*clientStream).doRequest(0x14000764600, 0x0?, 0x0?) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1431 golang.org/x/net/http2.(*ClientConn).roundTrip in goroutine 71 /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1336 goroutine 240 [IO wait] internal/poll.runtime_pollWait(0x1146fc000, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140007306c0?, 0x14000d10000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140007306c0, {0x14000d10000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000554a58, {0x14000d10000?, 0x14000983da8?, 0x104c7ba18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x106dc6d80, 0x140007545c0}, {0x106dc5ec8, 0x14000d0e000}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x106dc6d80, 0x140007545c0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x108261230?, {0x106dc6d80?, 0x140007545c0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x106dc6d80, 0x140007545c0}, {0x106dc5fc0, 0x14000554a58}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 239 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 241 [IO wait] internal/poll.runtime_pollWait(0x1146fb200, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000730780?, 0x14000a32000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000730780, {0x14000a32000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000554a70, {0x14000a32000?, 0x14000984da8?, 0x104c7ba18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x106dc6d80, 0x14000754600}, {0x106dc5ec8, 0x1400011c3f8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x106dc6d80, 0x14000754600}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x108261230?, {0x106dc6d80?, 0x14000754600?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x106dc6d80, 0x14000754600}, {0x106dc5fc0, 0x14000554a70}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 239 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 290 [select] os/exec.(*Cmd).watchCtx(0x140000dc480, 0x1400057ed20) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 239 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 243 [IO wait] internal/poll.runtime_pollWait(0x1146fbc00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000074d80?, 0x14000a2a000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000074d80, {0x14000a2a000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000428188, {0x14000a2a000?, 0x14001154da8?, 0x104c7ba18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x106dc6d80, 0x1400059c180}, {0x106dc5ec8, 0x1400011c3f0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x106dc6d80, 0x1400059c180}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x108261230?, {0x106dc6d80?, 0x1400059c180?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x106dc6d80, 0x1400059c180}, {0x106dc5fc0, 0x14000428188}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 238 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 245 [select] os/exec.(*Cmd).watchCtx(0x14000628180, 0x140006c6460) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 238 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 Interrupted by User First interrupt received; Ginkgo will run any cleanup and reporting nodes but will skip all remaining specs. Interrupt again to skip cleanup. Here's a current progress report: ImporterNetworkPolicy When resources are applied When virtual disks are applied (Spec Runtime: 5.665s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 In [It] (Node Runtime: 5.665s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 At [By Step] VDs should be in Ready phases (Step Runtime: 5.665s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 Spec Goroutine goroutine 234 [sync.WaitGroup.Wait] sync.runtime_SemacquireWaitGroup(0x104cb8b30?, 0x10?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000a18da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14001082520, 0x2, 0x106731e58?}, {0x10644a204, 0x28}, {{0x0, 0x0, 0x0}, 0x140005fea80, {0x14000057a10, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x10644a204, 0x28}, {{0x0, 0x0, 0x0}, 0x140005fea80, {0x14000057a10, 0x30}, {0x14000055900, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x10644a204, 0x28}, {0x106408eef, 0x5}, {{0x0, 0x0, 0x0}, 0x140005fea80, {0x14000057a10, 0x30}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func7.7({0x10640795c?, 0x0?}, {0x10644a204, 0x28}, {0x106408eef, 0x5}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 | func(resourceShortName string, resource kc.Resource, phase string) { | By(fmt.Sprintf("%ss should be in %s phases", resourceShortName, phase)) > WaitPhaseByLabel(resource, phase, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, reflect.Value.call({0x1069ae7e0?, 0x14000404e70?, 0x140000b4e18?}, {0x106408530, 0x4}, {0x140002c81e0, 0x3, 0x1050e11b0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/reflect/value.go:581 reflect.Value.Call({0x1069ae7e0?, 0x14000404e70?, 0x106e18ca0?}, {0x140002c81e0?, 0x0?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/reflect/value.go:365 > github.com/onsi/ginkgo/v2.invokeFunction({0x1069ae7e0, 0x14000404e70}, {0x1400061e800, 0x3, 0x14000771000?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/table_dsl.go:339 | } | > return reflect.ValueOf(function).Call(inValues) | } | > github.com/onsi/ginkgo/v2.generateTable.func2.2() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/table_dsl.go:303 | panic(err) | } > invokeFunction(internalBody, entry.parameters) | }) | } > github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x14000653c50?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 | | body := arg.(func()) > return func(SpecContext) { body() }, false | } | > github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 | }() | > node.Body(sc) | finished = true | }() > github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 | failureC := make(chan types.Failure) | > go func() { | finished := false | defer func() { Goroutines of Interest goroutine 28 [select, 2 minutes] > github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal.func1() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:32 | go func() { | for { > select { | case <-signalChannel: | handler() > github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:30 | } | ctx, cancel := context.WithCancel(context.Background()) > go func() { | for { | select { goroutine 100 [sync.Cond.Wait] sync.runtime_notifyListWait(0x14000764648, 0x1c) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x14000764638) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x14000764630, {0x1400099c000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400099a150?}, {0x1400099c000?, 0x1068a1b40?, 0x140010b9080?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140000a2ae0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140000a2ae0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x140007d6008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x140007d6008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x140007d6008, {0x14000b060db, 0xf9f25, 0x104dfdc74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x14000329348, {0x14000b060db, 0xf9f25, 0xf9f25}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000b060db?, 0x1400063d7c8?, 0x14000b05e61?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x140009c9da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140004323c0, {0x114708798, 0x1400099a030}, {0x106dc75a0, 0x140000753e0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 71 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 39 [select] > github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts.func2(0x0?) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:131 | var interruptCause InterruptCause | for { > select { | case <-signalChannel: | interruptCause = InterruptCauseSignal > github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:128 | } | > go func(abortChannel chan interface{}) { | var interruptCause InterruptCause | for { goroutine 102 [sync.Cond.Wait] sync.runtime_notifyListWait(0x140000dcf48, 0xa0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140000dcf38) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140000dcf30, {0x140005c4000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400099a150?}, {0x140005c4000?, 0x1068a1b40?, 0x1400070d550?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x14000075f20) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x14000075f20) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000aa0008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000aa0008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000aa0008, {0x14000c19a84, 0xe657c, 0x104dfdc74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400026c008, {0x14000c19a84, 0xe657c, 0xe657c}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000c19a84?, 0x140004ebdf8?, 0x14000c19963?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x14000033da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140004323c0, {0x114708798, 0x1400099af60}, {0x106dc75a0, 0x140000753e0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 71 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 239 [syscall] syscall.syscall6(0x100a18de0?, 0x11465e308?, 0x109028a78?, 0x90?, 0x140004a1808?, 0x1400028e2d0?, 0x14000982a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000982a88?, 0x104d3a6fc?, 0x90?, 0x106d36120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400019d110?, 0x14000982ac4, 0x14000a18de0?, 0x1400019d0a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000754640) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140000dbc08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140000dc480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140000dc480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a3490?, 0x106dcde88?, 0x1400019cfc0?}}, {0x106dee488?, 0x1400019cfc0?}, {0x14000564240?, 0xc232cad5b38f99a8?}, {0x106dc6d80, 0x140007545c0}, {0x106dc6d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a3490?, 0x1082db728?, 0x1082b1fa0?}}, {0x106dee488, 0x1400019cfc0}, {0x14000564240, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106df2a38?, 0x140004bd350?}, {0x10640bfba?, 0x25?}}, {0x10644a204, 0x28}, {0x14000a05901, 0x14}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 234 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 238 [syscall] syscall.syscall6(0x1009147b0?, 0x11479d048?, 0x109028108?, 0x90?, 0x140000a5008?, 0x140010d21b0?, 0x140000b4a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140000b4a88?, 0x104d3a6fc?, 0x90?, 0x106d36120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029c540?, 0x140000b4ac4, 0x140009147b0?, 0x1400029c4d0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x1400059c200) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140010d5408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000628180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000628180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a3490?, 0x106dcde88?, 0x1400029c3f0?}}, {0x106dee488?, 0x1400029c3f0?}, {0x140010fe000?, 0xc232cad5b38fd440?}, {0x106dc6d80, 0x1400059c180}, {0x106dc6d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a3490?, 0x1082db728?, 0x1082b1fa0?}}, {0x106dee488, 0x1400029c3f0}, {0x140010fe000, 0xbb}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106df2a38?, 0x140004bd350?}, {0x10640bfba?, 0x140000b4e88?}}, {0x10644a204, 0x28}, {0x14000a058f0, 0x10}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 234 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/11/25 22:28:31.476 (5.672s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 22:28:31.476 The list of pods is empty; nothing to dump. < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 22:28:37.775 (6.299s) > Enter [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/11/25 22:28:37.775 STEP: Delete manifests - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:50 @ 10/11/25 22:28:37.775 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/11/25 22:28:37.776 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/11/25 22:28:40.173 (2.398s) < Exit [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/11/25 22:28:40.173 (2.398s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x140000de828>: the container "virtualization-controller" was not found: virtualization-controller-6b445c5fbd-wckkk { errs: [ <*errors.errorString | 0x14000651570>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-6b445c5fbd-wckkk", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/11/25 22:28:40.586 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/11/25 22:28:40.185 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/11/25 22:28:40.185 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/11/25 22:28:40.185 [FAILED] Expected success, but got an error: <*errors.joinError | 0x140000de828>: the container "virtualization-controller" was not found: virtualization-controller-6b445c5fbd-wckkk { errs: [ <*errors.errorString | 0x14000651570>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-6b445c5fbd-wckkk", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/11/25 22:28:40.586 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/11/25 22:28:40.586 (401ms) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/11/25 22:28:40.586 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/11/25 22:28:44.488 (3.902s) + + + \ No newline at end of file diff --git a/artifacts/fixed-sds-20251011-172723-24071/junit.xml b/artifacts/fixed-sds-20251011-172723-24071/junit.xml new file mode 100644 index 0000000000..2899c78da6 --- /dev/null +++ b/artifacts/fixed-sds-20251011-172723-24071/junit.xml @@ -0,0 +1,672 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/11/25 18:36:29.726 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/11/25 18:36:34.63 (4.9s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/11/25 18:36:34.63 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/11/25 18:36:34.63 (0s) + + + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/11/25 18:36:34.634 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/11/25 18:36:36.626 (1.991s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/11/25 18:36:36.627 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/11/25 18:36:41.167 (4.538s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/11/25 18:36:41.167 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/11/25 18:36:41.167 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/11/25 18:36:41.168 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:97 @ 10/11/25 18:36:41.168 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/11/25 19:26:32.428 (34.583s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/11/25 19:26:32.428 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/11/25 19:26:32.428 (0s) + + + [INTERRUPTED] Interrupted by User In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/11/25 21:13:13.636 This is the Progress Report generated when the interrupt was received: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 4m35.212s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [It] (Node Runtime: 4m35.212s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 At [By Step] VDs should be in Ready phase (Step Runtime: 4m35.212s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 Spec Goroutine goroutine 78 [sync.WaitGroup.Wait, 5 minutes] sync.runtime_SemacquireWaitGroup(0x102104b30?, 0xe0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000900dc0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140007ac8e0, 0x2, 0x103b7de58?}, {0x103896204, 0x28}, {{0x0, 0x0, 0x0}, 0x140000351a0, {0x14000d0a180, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x103896204, 0x28}, {{0x0, 0x0, 0x0}, 0x140000351a0, {0x14000d0a180, 0x25}, {0x14000167660, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x103896204, 0x28}, {0x103854eef, 0x5}, {{0x0, 0x0, 0x0}, 0x140000351a0, {0x14000d0a180, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 6 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 169 [syscall, 5 minutes] syscall.syscall6(0x1006c0dd0?, 0x111ad8168?, 0x106474108?, 0x90?, 0x14000580008?, 0x14000932240?, 0x14000571a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000571a88?, 0x1021866fc?, 0x90?, 0x104182120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000270690?, 0x14000571ac4, 0x140006c0dd0?, 0x14000270620?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000287680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000a55408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x1400093a180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x1400093a180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x104219e88?, 0x14000270540?}}, {0x10423a488?, 0x14000270540?}, {0x140003463c0?, 0xc232c02d7558b148?}, {0x104212d80, 0x14000286600}, {0x104212d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x105727728?, 0x1056fdfa0?}}, {0x10423a488, 0x14000270540}, {0x140003463c0, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10423ea38?, 0x140004a1c20?}, {0x103857fba?, 0x14000571e88?}}, {0x103896204, 0x28}, {0x14000431480, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 78 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 170 [syscall, 5 minutes] syscall.syscall6(0x100900e00?, 0x10655b060?, 0x1064745c0?, 0x90?, 0x14000273808?, 0x140002085a0?, 0x14000948a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000948a88?, 0x1021866fc?, 0x90?, 0x104182120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002890a0?, 0x14000948ac4, 0x14000900e00?, 0x14000289030?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140006d64c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000bd6808?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140005ae480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140005ae480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x104219e88?, 0x14000288f50?}}, {0x10423a488?, 0x14000288f50?}, {0x140006f0480?, 0xc232c02d7558ad60?}, {0x104212d80, 0x140006d6440}, {0x104212d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x105727728?, 0x1056fdfa0?}}, {0x10423a488, 0x14000288f50}, {0x140006f0480, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10423ea38?, 0x140004a1c20?}, {0x103857fba?, 0x2d7375656874656d?}}, {0x103896204, 0x28}, {0x140004314a0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 78 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) Other Goroutines goroutine 6 [running] github.com/onsi/ginkgo/v2/internal.extractRunningGoroutines() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:181 github.com/onsi/ginkgo/v2/internal.NewProgressReport(_, {{0x140005ac080, 0x2, 0x2}, {0x140002700e0, 0x2, 0x2}, {0x14000280a80, 0x2, 0x2}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:75 github.com/onsi/ginkgo/v2/internal.(*Suite).generateProgressReport(_, _) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:381 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode(_, {0xdc, 0x4, {0x103868e1e, 0x11}, 0x140002fb610, {{0x1049418e1, 0x55}, 0x6b, {0x0, ...}, ...}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:980 github.com/onsi/ginkgo/v2/internal.(*group).attemptSpec(0x140004ff7f0, 0x1, {{0x140006b7208?, 0x140002700e0?, 0x2?}, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:199 github.com/onsi/ginkgo/v2/internal.(*group).run(0x140004ff7f0, {0x140008c8fc8?, 0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:349 github.com/onsi/ginkgo/v2/internal.(*Suite).runSpecs(0x14000021508, {0x103854f2b, 0x5}, {0x105727728, 0x0, 0x0}, {0x1400005a0f4, 0x3d}, 0x0, {0x14000092c08, ...}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:489 github.com/onsi/ginkgo/v2/internal.(*Suite).Run(_, {_, _}, {_, _, _}, {_, _}, _, {_, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:130 github.com/onsi/ginkgo/v2.RunSpecs({0x1042120a0, 0x14000583340}, {0x103854f2b, 0x5}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:300 github.com/deckhouse/virtualization/tests/e2e.TestE2E(0x14000583340) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:76 testing.tRunner(0x14000583340, 0x104201c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.(*T).Run in goroutine 1 /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1997 goroutine 1 [chan receive, 5 minutes] testing.(*T).Run(0x14000583180, {0x10385799a?, 0x14000517b38?}, 0x104201c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2005 testing.runTests.func1(0x14000583180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2477 testing.tRunner(0x14000583180, 0x14000517c68) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.runTests(0x140004a0c00, {0x1056acfd0, 0x1, 0x1}, {0x140005ad0e0?, 0xf?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2475 testing.(*M).Run(0x140004e12c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2337 main.main() _testmain.go:45 goroutine 32 [select, 5 minutes] github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal.func1() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:32 github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal in goroutine 6 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:30 goroutine 178 [IO wait, 5 minutes] internal/poll.runtime_pollWait(0x111cd9a00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x1400060e660?, 0x14000f98000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x1400060e660, {0x14000f98000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140007000b0, {0x14000f98000?, 0x1400045cda8?, 0x1020c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x104212d80, 0x14000286600}, {0x104211ec8, 0x140000a2020}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1400011b001?, {0x104212d80, 0x14000286600}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1056ad230?, {0x104212d80?, 0x14000286600?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x104212d80, 0x14000286600}, {0x104211fc0, 0x140007000b0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400093a000?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 169 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 70 [select] github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts.func2(0x104242480?) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:131 github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts in goroutine 6 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:128 goroutine 82 [syscall] os/signal.signal_recv() /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sigqueue.go:149 os/signal.loop() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal_unix.go:23 os/signal.Notify.func1.1 in goroutine 6 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal.go:152 goroutine 179 [IO wait, 3 minutes] internal/poll.runtime_pollWait(0x111cd9800, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x1400060e720?, 0x14000fa0000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x1400060e720, {0x14000fa0000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140007000c8, {0x14000fa0000?, 0x129?, 0x8000?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x104212d80, 0x14000287600}, {0x104211ec8, 0x140000a2028}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x140000b3e01?, {0x104212d80, 0x14000287600}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1056ad230?, {0x104212d80?, 0x14000287600?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x104212d80, 0x14000287600}, {0x104211fc0, 0x140007000c8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400011aee0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 169 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 173 [select, 5 minutes] os/exec.(*Cmd).watchCtx(0x140005ae480, 0x1400057e9a0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 170 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 172 [IO wait, 3 minutes] internal/poll.runtime_pollWait(0x111cd9e00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x1400069c6c0?, 0x14000f90000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x1400069c6c0, {0x14000f90000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000592a98, {0x14000f90000?, 0x129?, 0x8000?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x104212d80, 0x140006d6480}, {0x104211ec8, 0x140000a2018}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1400057e701?, {0x104212d80, 0x140006d6480}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1056ad230?, {0x104212d80?, 0x140006d6480?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x104212d80, 0x140006d6480}, {0x104211fc0, 0x14000592a98}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140005ae300?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 170 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 171 [IO wait, 5 minutes] internal/poll.runtime_pollWait(0x111cda200, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x1400069c600?, 0x14000f1e000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x1400069c600, {0x14000f1e000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000592a80, {0x14000f1e000?, 0x14000949da8?, 0x1020c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x104212d80, 0x140006d6440}, {0x104211ec8, 0x140001260c8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x3a2270756f726701?, {0x104212d80, 0x140006d6440}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1056ad230?, {0x104212d80?, 0x140006d6440?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x104212d80, 0x140006d6440}, {0x104211fc0, 0x14000592a80}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x2265746164707522?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 170 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 180 [select, 5 minutes] os/exec.(*Cmd).watchCtx(0x1400093a180, 0x1400011b1f0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 169 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 + > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/11/25 19:26:32.428 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 @ 10/11/25 19:26:32.428 [INTERRUPTED] Interrupted by User In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/11/25 21:13:13.636 This is the Progress Report generated when the interrupt was received: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 4m35.212s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [It] (Node Runtime: 4m35.212s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 At [By Step] VDs should be in Ready phase (Step Runtime: 4m35.212s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 Spec Goroutine goroutine 78 [sync.WaitGroup.Wait, 5 minutes] sync.runtime_SemacquireWaitGroup(0x102104b30?, 0xe0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000900dc0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140007ac8e0, 0x2, 0x103b7de58?}, {0x103896204, 0x28}, {{0x0, 0x0, 0x0}, 0x140000351a0, {0x14000d0a180, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x103896204, 0x28}, {{0x0, 0x0, 0x0}, 0x140000351a0, {0x14000d0a180, 0x25}, {0x14000167660, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x103896204, 0x28}, {0x103854eef, 0x5}, {{0x0, 0x0, 0x0}, 0x140000351a0, {0x14000d0a180, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 6 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 169 [syscall, 5 minutes] syscall.syscall6(0x1006c0dd0?, 0x111ad8168?, 0x106474108?, 0x90?, 0x14000580008?, 0x14000932240?, 0x14000571a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000571a88?, 0x1021866fc?, 0x90?, 0x104182120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000270690?, 0x14000571ac4, 0x140006c0dd0?, 0x14000270620?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000287680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000a55408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x1400093a180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x1400093a180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x104219e88?, 0x14000270540?}}, {0x10423a488?, 0x14000270540?}, {0x140003463c0?, 0xc232c02d7558b148?}, {0x104212d80, 0x14000286600}, {0x104212d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x105727728?, 0x1056fdfa0?}}, {0x10423a488, 0x14000270540}, {0x140003463c0, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10423ea38?, 0x140004a1c20?}, {0x103857fba?, 0x14000571e88?}}, {0x103896204, 0x28}, {0x14000431480, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 78 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 170 [syscall, 5 minutes] syscall.syscall6(0x100900e00?, 0x10655b060?, 0x1064745c0?, 0x90?, 0x14000273808?, 0x140002085a0?, 0x14000948a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000948a88?, 0x1021866fc?, 0x90?, 0x104182120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002890a0?, 0x14000948ac4, 0x14000900e00?, 0x14000289030?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140006d64c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000bd6808?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140005ae480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140005ae480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x104219e88?, 0x14000288f50?}}, {0x10423a488?, 0x14000288f50?}, {0x140006f0480?, 0xc232c02d7558ad60?}, {0x104212d80, 0x140006d6440}, {0x104212d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x105727728?, 0x1056fdfa0?}}, {0x10423a488, 0x14000288f50}, {0x140006f0480, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10423ea38?, 0x140004a1c20?}, {0x103857fba?, 0x2d7375656874656d?}}, {0x103896204, 0x28}, {0x140004314a0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 78 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) Other Goroutines goroutine 6 [running] github.com/onsi/ginkgo/v2/internal.extractRunningGoroutines() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:181 github.com/onsi/ginkgo/v2/internal.NewProgressReport(_, {{0x140005ac080, 0x2, 0x2}, {0x140002700e0, 0x2, 0x2}, {0x14000280a80, 0x2, 0x2}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:75 github.com/onsi/ginkgo/v2/internal.(*Suite).generateProgressReport(_, _) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:381 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode(_, {0xdc, 0x4, {0x103868e1e, 0x11}, 0x140002fb610, {{0x1049418e1, 0x55}, 0x6b, {0x0, ...}, ...}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:980 github.com/onsi/ginkgo/v2/internal.(*group).attemptSpec(0x140004ff7f0, 0x1, {{0x140006b7208?, 0x140002700e0?, 0x2?}, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:199 github.com/onsi/ginkgo/v2/internal.(*group).run(0x140004ff7f0, {0x140008c8fc8?, 0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:349 github.com/onsi/ginkgo/v2/internal.(*Suite).runSpecs(0x14000021508, {0x103854f2b, 0x5}, {0x105727728, 0x0, 0x0}, {0x1400005a0f4, 0x3d}, 0x0, {0x14000092c08, ...}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:489 github.com/onsi/ginkgo/v2/internal.(*Suite).Run(_, {_, _}, {_, _, _}, {_, _}, _, {_, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:130 github.com/onsi/ginkgo/v2.RunSpecs({0x1042120a0, 0x14000583340}, {0x103854f2b, 0x5}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:300 github.com/deckhouse/virtualization/tests/e2e.TestE2E(0x14000583340) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:76 testing.tRunner(0x14000583340, 0x104201c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.(*T).Run in goroutine 1 /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1997 goroutine 1 [chan receive, 5 minutes] testing.(*T).Run(0x14000583180, {0x10385799a?, 0x14000517b38?}, 0x104201c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2005 testing.runTests.func1(0x14000583180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2477 testing.tRunner(0x14000583180, 0x14000517c68) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.runTests(0x140004a0c00, {0x1056acfd0, 0x1, 0x1}, {0x140005ad0e0?, 0xf?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2475 testing.(*M).Run(0x140004e12c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2337 main.main() _testmain.go:45 goroutine 32 [select, 5 minutes] github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal.func1() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:32 github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal in goroutine 6 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:30 goroutine 178 [IO wait, 5 minutes] internal/poll.runtime_pollWait(0x111cd9a00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x1400060e660?, 0x14000f98000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x1400060e660, {0x14000f98000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140007000b0, {0x14000f98000?, 0x1400045cda8?, 0x1020c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x104212d80, 0x14000286600}, {0x104211ec8, 0x140000a2020}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1400011b001?, {0x104212d80, 0x14000286600}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1056ad230?, {0x104212d80?, 0x14000286600?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x104212d80, 0x14000286600}, {0x104211fc0, 0x140007000b0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400093a000?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 169 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 70 [select] github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts.func2(0x104242480?) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:131 github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts in goroutine 6 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:128 goroutine 82 [syscall] os/signal.signal_recv() /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sigqueue.go:149 os/signal.loop() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal_unix.go:23 os/signal.Notify.func1.1 in goroutine 6 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal.go:152 goroutine 179 [IO wait, 3 minutes] internal/poll.runtime_pollWait(0x111cd9800, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x1400060e720?, 0x14000fa0000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x1400060e720, {0x14000fa0000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140007000c8, {0x14000fa0000?, 0x129?, 0x8000?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x104212d80, 0x14000287600}, {0x104211ec8, 0x140000a2028}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x140000b3e01?, {0x104212d80, 0x14000287600}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1056ad230?, {0x104212d80?, 0x14000287600?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x104212d80, 0x14000287600}, {0x104211fc0, 0x140007000c8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400011aee0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 169 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 173 [select, 5 minutes] os/exec.(*Cmd).watchCtx(0x140005ae480, 0x1400057e9a0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 170 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 172 [IO wait, 3 minutes] internal/poll.runtime_pollWait(0x111cd9e00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x1400069c6c0?, 0x14000f90000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x1400069c6c0, {0x14000f90000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000592a98, {0x14000f90000?, 0x129?, 0x8000?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x104212d80, 0x140006d6480}, {0x104211ec8, 0x140000a2018}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1400057e701?, {0x104212d80, 0x140006d6480}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1056ad230?, {0x104212d80?, 0x140006d6480?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x104212d80, 0x140006d6480}, {0x104211fc0, 0x14000592a98}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140005ae300?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 170 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 171 [IO wait, 5 minutes] internal/poll.runtime_pollWait(0x111cda200, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x1400069c600?, 0x14000f1e000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x1400069c600, {0x14000f1e000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000592a80, {0x14000f1e000?, 0x14000949da8?, 0x1020c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x104212d80, 0x140006d6440}, {0x104211ec8, 0x140001260c8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x3a2270756f726701?, {0x104212d80, 0x140006d6440}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1056ad230?, {0x104212d80?, 0x140006d6440?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x104212d80, 0x140006d6440}, {0x104211fc0, 0x14000592a80}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x2265746164707522?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 170 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 180 [select, 5 minutes] os/exec.(*Cmd).watchCtx(0x1400093a180, 0x1400011b1f0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 169 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 Interrupted by User First interrupt received; Ginkgo will run any cleanup and reporting nodes but will skip all remaining specs. Interrupt again to skip cleanup. Here's a current progress report: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 4m35.212s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [It] (Node Runtime: 4m35.212s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 At [By Step] VDs should be in Ready phase (Step Runtime: 4m35.212s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 Spec Goroutine goroutine 78 [sync.WaitGroup.Wait, 5 minutes] sync.runtime_SemacquireWaitGroup(0x102104b30?, 0xe0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000900dc0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140007ac8e0, 0x2, 0x103b7de58?}, {0x103896204, 0x28}, {{0x0, 0x0, 0x0}, 0x140000351a0, {0x14000d0a180, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x103896204, 0x28}, {{0x0, 0x0, 0x0}, 0x140000351a0, {0x14000d0a180, 0x25}, {0x14000167660, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x103896204, 0x28}, {0x103854eef, 0x5}, {{0x0, 0x0, 0x0}, 0x140000351a0, {0x14000d0a180, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 6 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 169 [syscall, 5 minutes] syscall.syscall6(0x1006c0dd0?, 0x111ad8168?, 0x106474108?, 0x90?, 0x14000580008?, 0x14000932240?, 0x14000571a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000571a88?, 0x1021866fc?, 0x90?, 0x104182120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000270690?, 0x14000571ac4, 0x140006c0dd0?, 0x14000270620?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000287680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000a55408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x1400093a180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x1400093a180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x104219e88?, 0x14000270540?}}, {0x10423a488?, 0x14000270540?}, {0x140003463c0?, 0xc232c02d7558b148?}, {0x104212d80, 0x14000286600}, {0x104212d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x105727728?, 0x1056fdfa0?}}, {0x10423a488, 0x14000270540}, {0x140003463c0, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10423ea38?, 0x140004a1c20?}, {0x103857fba?, 0x14000571e88?}}, {0x103896204, 0x28}, {0x14000431480, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 78 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 170 [syscall, 5 minutes] syscall.syscall6(0x100900e00?, 0x10655b060?, 0x1064745c0?, 0x90?, 0x14000273808?, 0x140002085a0?, 0x14000948a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000948a88?, 0x1021866fc?, 0x90?, 0x104182120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002890a0?, 0x14000948ac4, 0x14000900e00?, 0x14000289030?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140006d64c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000bd6808?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140005ae480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140005ae480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x104219e88?, 0x14000288f50?}}, {0x10423a488?, 0x14000288f50?}, {0x140006f0480?, 0xc232c02d7558ad60?}, {0x104212d80, 0x140006d6440}, {0x104212d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x105727728?, 0x1056fdfa0?}}, {0x10423a488, 0x14000288f50}, {0x140006f0480, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10423ea38?, 0x140004a1c20?}, {0x103857fba?, 0x2d7375656874656d?}}, {0x103896204, 0x28}, {0x140004314a0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 78 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/11/25 21:13:13.64 (4m35.215s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/11/25 21:13:13.64 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/11/25 21:13:24.482 (10.843s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x14000a7e390>: the container "virtualization-controller" was not found: virtualization-controller-86578b98f6-t2wr5 the container "virtualization-controller" was restarted: virtualization-controller-9c76787d-vv5ng { errs: [ <*errors.joinError | 0x14000a7e378>{ errs: [ <*errors.errorString | 0x140002a29f0>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-86578b98f6-t2wr5", }, ], }, <*errors.errorString | 0x140002a2aa0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-9c76787d-vv5ng", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/11/25 21:13:25.507 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/11/25 21:13:24.487 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/11/25 21:13:24.487 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/11/25 21:13:24.487 [FAILED] Expected success, but got an error: <*errors.joinError | 0x14000a7e390>: the container "virtualization-controller" was not found: virtualization-controller-86578b98f6-t2wr5 the container "virtualization-controller" was restarted: virtualization-controller-9c76787d-vv5ng { errs: [ <*errors.joinError | 0x14000a7e378>{ errs: [ <*errors.errorString | 0x140002a29f0>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-86578b98f6-t2wr5", }, ], }, <*errors.errorString | 0x140002a2aa0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-9c76787d-vv5ng", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/11/25 21:13:25.507 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/11/25 21:13:25.507 (1.02s) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/11/25 21:13:25.507 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/11/25 21:13:29.153 (3.646s) + + + \ No newline at end of file diff --git a/artifacts/fixed-sds-20251011-211703-12576/junit.xml b/artifacts/fixed-sds-20251011-211703-12576/junit.xml new file mode 100644 index 0000000000..afac413702 --- /dev/null +++ b/artifacts/fixed-sds-20251011-211703-12576/junit.xml @@ -0,0 +1,673 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/11/25 21:34:52.839 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/11/25 21:34:57.578 (4.739s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/11/25 21:34:57.578 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/11/25 21:34:57.578 (0s) + + + [INTERRUPTED] Interrupted by User In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/11/25 22:28:31.47 This is the Progress Report generated when the interrupt was received: VirtualMachineEvacuation Evacuation (Spec Runtime: 8m9.737s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 In [It] (Node Runtime: 8m1.49s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 At [By Step] Virtual machine agents should be ready (Step Runtime: 8m1.49s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 Spec Goroutine goroutine 146 [sync.WaitGroup.Wait, 8 minutes] sync.runtime_SemacquireWaitGroup(0x100200b30?, 0x60?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x140005bd110) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140001aa060, 0x2, 0x101c79e58?}, {0x101996a2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400067b320, {0x1400040e720, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x101996a2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400067b320, {0x1400040e720, 0x26}, {0x1400040ef00, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x101996a2d, 0x2a}, {0x101953a0a, 0x7}, {{0x0, 0x0, 0x0}, 0x1400067b320, {0x1400040e720, 0x26}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x1400067b320, {0x1400040e720, 0x26}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func17.5() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 | It("Evacuation", func() { | By("Virtual machine agents should be ready") > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 150 [syscall, 8 minutes] syscall.syscall6(0x100010990?, 0x12de212c0?, 0x104570a78?, 0x90?, 0x14000980008?, 0x14000888090?, 0x14000561a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000561a88?, 0x1002826fc?, 0x90?, 0x10227e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000312700?, 0x14000561ac4, 0x14000010990?, 0x14000312540?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140004c42c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140008a6408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140008a4000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140008a4000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140003319e0?, 0x102315e88?, 0x140003120e0?}}, {0x102336488?, 0x140003120e0?}, {0x140008a2000?, 0xc232c7b5b94f8240?}, {0x10230ed80, 0x140004c4000}, {0x10230ed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140003319e0?, 0x103823728?, 0x1037f9fa0?}}, {0x102336488, 0x140003120e0}, {0x140008a2000, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10233aa38?, 0x14000627e78?}, {0x101953fba?, 0x1400019fe98?}}, {0x101996a2d, 0x2a}, {0x1400059ca00, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 146 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 151 [syscall, 8 minutes] syscall.syscall6(0x1005bd150?, 0x12de10a00?, 0x1045705c0?, 0x90?, 0x14000a00008?, 0x14000916240?, 0x14000899a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000899a88?, 0x1002826fc?, 0x90?, 0x10227e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400031eaf0?, 0x14000899ac4, 0x140005bd150?, 0x1400031ea80?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000487680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000decc08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140005a4480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140005a4480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140003319e0?, 0x102315e88?, 0x1400031e9a0?}}, {0x102336488?, 0x1400031e9a0?}, {0x14000df0000?, 0xc232c7b5b94f5b30?}, {0x10230ed80, 0x14000487580}, {0x10230ed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140003319e0?, 0x103823728?, 0x1037f9fa0?}}, {0x102336488, 0x1400031e9a0}, {0x14000df0000, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10233aa38?, 0x14000627e78?}, {0x101953fba?, 0x1400009fe68?}}, {0x101996a2d, 0x2a}, {0x1400059ca20, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 146 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) Other Goroutines goroutine 8 [running] github.com/onsi/ginkgo/v2/internal.extractRunningGoroutines() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:181 github.com/onsi/ginkgo/v2/internal.NewProgressReport(_, {{0x140006be000, 0x1, 0x1}, {0x14000486c00, 0x1, 0x1}, {0x14000882048, 0x1, 0x1}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:75 github.com/onsi/ginkgo/v2/internal.(*Suite).generateProgressReport(_, _) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:381 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode(_, {0x147, 0x4, {0x10195889b, 0xa}, 0x140007957c0, {{0x102a3db2f, 0x53}, 0x60, {0x0, ...}, ...}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:980 github.com/onsi/ginkgo/v2/internal.(*group).attemptSpec(0x14000ddb7f0, 0x1, {{0x140005e8008?, 0x14000486c00?, 0x1?}, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:199 github.com/onsi/ginkgo/v2/internal.(*group).run(0x14000ddb7f0, {0x140003d2e40?, 0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:349 github.com/onsi/ginkgo/v2/internal.(*Suite).runSpecs(0x14000029508, {0x101950f2b, 0x5}, {0x103823728, 0x0, 0x0}, {0x1400005a0f4, 0x3d}, 0x0, {0x14000092c08, ...}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:489 github.com/onsi/ginkgo/v2/internal.(*Suite).Run(_, {_, _}, {_, _, _}, {_, _}, _, {_, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:130 github.com/onsi/ginkgo/v2.RunSpecs({0x10230e0a0, 0x14000582a80}, {0x101950f2b, 0x5}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:300 github.com/deckhouse/virtualization/tests/e2e.TestE2E(0x14000582a80) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:76 testing.tRunner(0x14000582a80, 0x1022fdc50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.(*T).Run in goroutine 1 /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1997 goroutine 1 [chan receive, 8 minutes] testing.(*T).Run(0x140005828c0, {0x10195399a?, 0x14000599b38?}, 0x1022fdc50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2005 testing.runTests.func1(0x140005828c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2477 testing.tRunner(0x140005828c0, 0x14000599c68) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.runTests(0x14000626d80, {0x1037a8fd0, 0x1, 0x1}, {0x140003295c0?, 0xf?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2475 testing.(*M).Run(0x140005061e0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2337 main.main() _testmain.go:45 goroutine 56 [syscall] os/signal.signal_recv() /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sigqueue.go:149 os/signal.loop() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal_unix.go:23 os/signal.Notify.func1.1 in goroutine 8 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal.go:152 goroutine 57 [select] github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts.func2(0x0?) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:131 github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:128 goroutine 58 [select, 8 minutes] github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal.func1() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:32 github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:30 goroutine 163 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12ddf8400, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000118360?, 0x14000998000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000118360, {0x14000998000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000128048, {0x14000998000?, 0x140001995a8?, 0x1001c3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10230ed80, 0x140004c4200}, {0x10230dec8, 0x14000612008}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x140000de401?, {0x10230ed80, 0x140004c4200}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037a9230?, {0x10230ed80?, 0x140004c4200?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10230ed80, 0x140004c4200}, {0x10230dfc0, 0x14000128048}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140005a4180?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 150 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 162 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12ddf8c00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140001181e0?, 0x14000900000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140001181e0, {0x14000900000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000128028, {0x14000900000?, 0x1400019fda8?, 0x1001c3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10230ed80, 0x140004c4000}, {0x10230dec8, 0x140000a3258}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000627e01?, {0x10230ed80, 0x140004c4000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037a9230?, {0x10230ed80?, 0x140004c4000?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10230ed80, 0x140004c4000}, {0x10230dfc0, 0x14000128028}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140005bd108?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 150 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 164 [select, 8 minutes] os/exec.(*Cmd).watchCtx(0x140008a4000, 0x140008c6070) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 150 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 152 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12ddf8800, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140007186c0?, 0x14000d60000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140007186c0, {0x14000d60000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140004c2228, {0x14000d60000?, 0x1400009fda8?, 0x1001c3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10230ed80, 0x14000487580}, {0x10230dec8, 0x140006b20a8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000627e01?, {0x10230ed80, 0x14000487580}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037a9230?, {0x10230ed80?, 0x14000487580?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10230ed80, 0x14000487580}, {0x10230dfc0, 0x140004c2228}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140005bd108?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 151 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 153 [IO wait, 4 minutes] internal/poll.runtime_pollWait(0x12ddf8600, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000718780?, 0x14000848000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000718780, {0x14000848000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140004c2250, {0x14000848000?, 0x129?, 0x8000?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10230ed80, 0x14000487600}, {0x10230dec8, 0x140000a3250}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x10230ed80, 0x14000487600}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037a9230?, {0x10230ed80?, 0x14000487600?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10230ed80, 0x14000487600}, {0x10230dfc0, 0x140004c2250}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140001829a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 151 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 154 [select, 8 minutes] os/exec.(*Cmd).watchCtx(0x140005a4480, 0x140000de850) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 151 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/11/25 21:34:57.58 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/11/25 21:34:59.105 (1.525s) > Enter [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/11/25 21:34:59.105 < Exit [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/11/25 21:35:05.827 (6.721s) > Enter [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/11/25 21:35:05.827 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 @ 10/11/25 21:35:05.827 [INTERRUPTED] Interrupted by User In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/11/25 22:28:31.47 This is the Progress Report generated when the interrupt was received: VirtualMachineEvacuation Evacuation (Spec Runtime: 8m9.737s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 In [It] (Node Runtime: 8m1.49s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 At [By Step] Virtual machine agents should be ready (Step Runtime: 8m1.49s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 Spec Goroutine goroutine 146 [sync.WaitGroup.Wait, 8 minutes] sync.runtime_SemacquireWaitGroup(0x100200b30?, 0x60?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x140005bd110) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140001aa060, 0x2, 0x101c79e58?}, {0x101996a2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400067b320, {0x1400040e720, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x101996a2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400067b320, {0x1400040e720, 0x26}, {0x1400040ef00, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x101996a2d, 0x2a}, {0x101953a0a, 0x7}, {{0x0, 0x0, 0x0}, 0x1400067b320, {0x1400040e720, 0x26}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x1400067b320, {0x1400040e720, 0x26}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func17.5() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 | It("Evacuation", func() { | By("Virtual machine agents should be ready") > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 150 [syscall, 8 minutes] syscall.syscall6(0x100010990?, 0x12de212c0?, 0x104570a78?, 0x90?, 0x14000980008?, 0x14000888090?, 0x14000561a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000561a88?, 0x1002826fc?, 0x90?, 0x10227e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000312700?, 0x14000561ac4, 0x14000010990?, 0x14000312540?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140004c42c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140008a6408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140008a4000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140008a4000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140003319e0?, 0x102315e88?, 0x140003120e0?}}, {0x102336488?, 0x140003120e0?}, {0x140008a2000?, 0xc232c7b5b94f8240?}, {0x10230ed80, 0x140004c4000}, {0x10230ed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140003319e0?, 0x103823728?, 0x1037f9fa0?}}, {0x102336488, 0x140003120e0}, {0x140008a2000, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10233aa38?, 0x14000627e78?}, {0x101953fba?, 0x1400019fe98?}}, {0x101996a2d, 0x2a}, {0x1400059ca00, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 146 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 151 [syscall, 8 minutes] syscall.syscall6(0x1005bd150?, 0x12de10a00?, 0x1045705c0?, 0x90?, 0x14000a00008?, 0x14000916240?, 0x14000899a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000899a88?, 0x1002826fc?, 0x90?, 0x10227e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400031eaf0?, 0x14000899ac4, 0x140005bd150?, 0x1400031ea80?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000487680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000decc08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140005a4480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140005a4480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140003319e0?, 0x102315e88?, 0x1400031e9a0?}}, {0x102336488?, 0x1400031e9a0?}, {0x14000df0000?, 0xc232c7b5b94f5b30?}, {0x10230ed80, 0x14000487580}, {0x10230ed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140003319e0?, 0x103823728?, 0x1037f9fa0?}}, {0x102336488, 0x1400031e9a0}, {0x14000df0000, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10233aa38?, 0x14000627e78?}, {0x101953fba?, 0x1400009fe68?}}, {0x101996a2d, 0x2a}, {0x1400059ca20, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 146 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) Other Goroutines goroutine 8 [running] github.com/onsi/ginkgo/v2/internal.extractRunningGoroutines() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:181 github.com/onsi/ginkgo/v2/internal.NewProgressReport(_, {{0x140006be000, 0x1, 0x1}, {0x14000486c00, 0x1, 0x1}, {0x14000882048, 0x1, 0x1}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:75 github.com/onsi/ginkgo/v2/internal.(*Suite).generateProgressReport(_, _) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:381 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode(_, {0x147, 0x4, {0x10195889b, 0xa}, 0x140007957c0, {{0x102a3db2f, 0x53}, 0x60, {0x0, ...}, ...}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:980 github.com/onsi/ginkgo/v2/internal.(*group).attemptSpec(0x14000ddb7f0, 0x1, {{0x140005e8008?, 0x14000486c00?, 0x1?}, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:199 github.com/onsi/ginkgo/v2/internal.(*group).run(0x14000ddb7f0, {0x140003d2e40?, 0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:349 github.com/onsi/ginkgo/v2/internal.(*Suite).runSpecs(0x14000029508, {0x101950f2b, 0x5}, {0x103823728, 0x0, 0x0}, {0x1400005a0f4, 0x3d}, 0x0, {0x14000092c08, ...}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:489 github.com/onsi/ginkgo/v2/internal.(*Suite).Run(_, {_, _}, {_, _, _}, {_, _}, _, {_, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:130 github.com/onsi/ginkgo/v2.RunSpecs({0x10230e0a0, 0x14000582a80}, {0x101950f2b, 0x5}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:300 github.com/deckhouse/virtualization/tests/e2e.TestE2E(0x14000582a80) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:76 testing.tRunner(0x14000582a80, 0x1022fdc50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.(*T).Run in goroutine 1 /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1997 goroutine 1 [chan receive, 8 minutes] testing.(*T).Run(0x140005828c0, {0x10195399a?, 0x14000599b38?}, 0x1022fdc50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2005 testing.runTests.func1(0x140005828c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2477 testing.tRunner(0x140005828c0, 0x14000599c68) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.runTests(0x14000626d80, {0x1037a8fd0, 0x1, 0x1}, {0x140003295c0?, 0xf?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2475 testing.(*M).Run(0x140005061e0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2337 main.main() _testmain.go:45 goroutine 56 [syscall] os/signal.signal_recv() /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sigqueue.go:149 os/signal.loop() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal_unix.go:23 os/signal.Notify.func1.1 in goroutine 8 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal.go:152 goroutine 57 [select] github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts.func2(0x0?) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:131 github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:128 goroutine 58 [select, 8 minutes] github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal.func1() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:32 github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:30 goroutine 163 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12ddf8400, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000118360?, 0x14000998000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000118360, {0x14000998000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000128048, {0x14000998000?, 0x140001995a8?, 0x1001c3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10230ed80, 0x140004c4200}, {0x10230dec8, 0x14000612008}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x140000de401?, {0x10230ed80, 0x140004c4200}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037a9230?, {0x10230ed80?, 0x140004c4200?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10230ed80, 0x140004c4200}, {0x10230dfc0, 0x14000128048}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140005a4180?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 150 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 162 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12ddf8c00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140001181e0?, 0x14000900000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140001181e0, {0x14000900000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000128028, {0x14000900000?, 0x1400019fda8?, 0x1001c3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10230ed80, 0x140004c4000}, {0x10230dec8, 0x140000a3258}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000627e01?, {0x10230ed80, 0x140004c4000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037a9230?, {0x10230ed80?, 0x140004c4000?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10230ed80, 0x140004c4000}, {0x10230dfc0, 0x14000128028}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140005bd108?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 150 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 164 [select, 8 minutes] os/exec.(*Cmd).watchCtx(0x140008a4000, 0x140008c6070) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 150 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 152 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12ddf8800, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140007186c0?, 0x14000d60000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140007186c0, {0x14000d60000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140004c2228, {0x14000d60000?, 0x1400009fda8?, 0x1001c3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10230ed80, 0x14000487580}, {0x10230dec8, 0x140006b20a8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000627e01?, {0x10230ed80, 0x14000487580}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037a9230?, {0x10230ed80?, 0x14000487580?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10230ed80, 0x14000487580}, {0x10230dfc0, 0x140004c2228}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140005bd108?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 151 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 153 [IO wait, 4 minutes] internal/poll.runtime_pollWait(0x12ddf8600, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000718780?, 0x14000848000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000718780, {0x14000848000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140004c2250, {0x14000848000?, 0x129?, 0x8000?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10230ed80, 0x14000487600}, {0x10230dec8, 0x140000a3250}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x10230ed80, 0x14000487600}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037a9230?, {0x10230ed80?, 0x14000487600?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10230ed80, 0x14000487600}, {0x10230dfc0, 0x140004c2250}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140001829a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 151 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 154 [select, 8 minutes] os/exec.(*Cmd).watchCtx(0x140005a4480, 0x140000de850) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 151 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 Interrupted by User First interrupt received; Ginkgo will run any cleanup and reporting nodes but will skip all remaining specs. Interrupt again to skip cleanup. Here's a current progress report: VirtualMachineEvacuation Evacuation (Spec Runtime: 8m9.737s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 In [It] (Node Runtime: 8m1.49s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 At [By Step] Virtual machine agents should be ready (Step Runtime: 8m1.49s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 Spec Goroutine goroutine 146 [sync.WaitGroup.Wait, 8 minutes] sync.runtime_SemacquireWaitGroup(0x100200b30?, 0x60?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x140005bd110) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140001aa060, 0x2, 0x101c79e58?}, {0x101996a2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400067b320, {0x1400040e720, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x101996a2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400067b320, {0x1400040e720, 0x26}, {0x1400040ef00, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x101996a2d, 0x2a}, {0x101953a0a, 0x7}, {{0x0, 0x0, 0x0}, 0x1400067b320, {0x1400040e720, 0x26}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x1400067b320, {0x1400040e720, 0x26}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func17.5() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 | It("Evacuation", func() { | By("Virtual machine agents should be ready") > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 150 [syscall, 8 minutes] syscall.syscall6(0x100010990?, 0x12de212c0?, 0x104570a78?, 0x90?, 0x14000980008?, 0x14000888090?, 0x14000561a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000561a88?, 0x1002826fc?, 0x90?, 0x10227e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000312700?, 0x14000561ac4, 0x14000010990?, 0x14000312540?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140004c42c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140008a6408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140008a4000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140008a4000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140003319e0?, 0x102315e88?, 0x140003120e0?}}, {0x102336488?, 0x140003120e0?}, {0x140008a2000?, 0xc232c7b5b94f8240?}, {0x10230ed80, 0x140004c4000}, {0x10230ed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140003319e0?, 0x103823728?, 0x1037f9fa0?}}, {0x102336488, 0x140003120e0}, {0x140008a2000, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10233aa38?, 0x14000627e78?}, {0x101953fba?, 0x1400019fe98?}}, {0x101996a2d, 0x2a}, {0x1400059ca00, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 146 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 151 [syscall, 8 minutes] syscall.syscall6(0x1005bd150?, 0x12de10a00?, 0x1045705c0?, 0x90?, 0x14000a00008?, 0x14000916240?, 0x14000899a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000899a88?, 0x1002826fc?, 0x90?, 0x10227e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400031eaf0?, 0x14000899ac4, 0x140005bd150?, 0x1400031ea80?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000487680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000decc08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140005a4480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140005a4480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140003319e0?, 0x102315e88?, 0x1400031e9a0?}}, {0x102336488?, 0x1400031e9a0?}, {0x14000df0000?, 0xc232c7b5b94f5b30?}, {0x10230ed80, 0x14000487580}, {0x10230ed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140003319e0?, 0x103823728?, 0x1037f9fa0?}}, {0x102336488, 0x1400031e9a0}, {0x14000df0000, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10233aa38?, 0x14000627e78?}, {0x101953fba?, 0x1400009fe68?}}, {0x101996a2d, 0x2a}, {0x1400059ca20, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 146 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/11/25 22:28:31.475 (8m1.495s) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/11/25 22:28:31.475 Failed to get logs: PodName: virt-launcher-head-5073ae15-vm-migration-bios-596s7 Error: Error from server (BadRequest): container "d8v-compute" in pod "virt-launcher-head-5073ae15-vm-migration-bios-596s7" is waiting to start: ContainerCreating STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/11/25 22:28:42.734 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/11/25 22:28:58.881 (16.147s) [FAILED] cannot delete test case resources cmd: kubectl delete virtualmachineoperations.virtualization.deckhouse.io --namespace head-5073ae15-end-to-end-vm-evacuation -l 'testcase=vm-evacuation' stderr: E1011 22:28:55.766456 51152 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server is currently unable to handle the request" E1011 22:28:56.738429 51152 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1011 22:28:56.996501 51152 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1011 22:28:57.229269 51152 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1011 22:28:57.472235 51152 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1011 22:28:57.707857 51152 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1011 22:28:57.940804 51152 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1011 22:28:58.173560 51152 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1011 22:28:58.404391 51152 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1011 22:28:58.634304 51152 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" E1011 22:28:58.874581 51152 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: the server could not find the requested resource" Error from server (NotFound): the server could not find the requested resource Unexpected error: <*exec.ExitError | 0x140005de040>: exit status 1 { ProcessState: { pid: 51152, status: 256, rusage: { Utime: { Sec: 0, Usec: 99422, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 92695, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 51970048, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3987, Majflt: 5, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 44, Msgrcv: 44, Nsignals: 77, Nvcsw: 32, Nivcsw: 2527, }, }, Stderr: nil, } occurred In [AfterEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:626 @ 10/11/25 22:28:58.881 < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/11/25 22:28:58.881 (27.407s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + [FAILED] Expected success, but got an error: <*errors.StatusError | 0x140008fe960>: the server could not find the requested resource (get pods) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "the server could not find the requested resource (get pods)", Reason: "NotFound", Details: { Name: "", Group: "", Kind: "pods", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>404 Not Found</title></head>\r\n<body>\r\n<center><h1>404 Not Found</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 404, }, } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/11/25 22:28:59.667 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/11/25 22:28:58.886 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/11/25 22:28:58.886 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/11/25 22:28:58.886 [FAILED] Expected success, but got an error: <*errors.StatusError | 0x140008fe960>: the server could not find the requested resource (get pods) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "the server could not find the requested resource (get pods)", Reason: "NotFound", Details: { Name: "", Group: "", Kind: "pods", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>404 Not Found</title></head>\r\n<body>\r\n<center><h1>404 Not Found</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 404, }, } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/11/25 22:28:59.667 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/11/25 22:28:59.667 (781ms) + + + [FAILED] Expected success, but got an error: <*errors.StatusError | 0x14000684320>: the server could not find the requested resource (get projects.deckhouse.io) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "the server could not find the requested resource (get projects.deckhouse.io)", Reason: "NotFound", Details: { Name: "", Group: "deckhouse.io", Kind: "projects", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>404 Not Found</title></head>\r\n<body>\r\n<center><h1>404 Not Found</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 404, }, } In [DeferCleanup (Suite)] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:177 @ 10/11/25 22:28:59.783 + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/11/25 22:28:59.667 [FAILED] Expected success, but got an error: <*errors.StatusError | 0x14000684320>: the server could not find the requested resource (get projects.deckhouse.io) { ErrStatus: { TypeMeta: {Kind: "", APIVersion: ""}, ListMeta: { SelfLink: "", ResourceVersion: "", Continue: "", RemainingItemCount: nil, }, Status: "Failure", Message: "the server could not find the requested resource (get projects.deckhouse.io)", Reason: "NotFound", Details: { Name: "", Group: "deckhouse.io", Kind: "projects", UID: "", Causes: [ { Type: "UnexpectedServerResponse", Message: "<html>\r\n<head><title>404 Not Found</title></head>\r\n<body>\r\n<center><h1>404 Not Found</h1></center>\r\n<hr><center>nginx</center>\r\n</body>\r\n</html>", Field: "", }, ], RetryAfterSeconds: 0, }, Code: 404, }, } In [DeferCleanup (Suite)] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:177 @ 10/11/25 22:28:59.783 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/11/25 22:28:59.783 (116ms) + + + \ No newline at end of file diff --git a/artifacts/m-cephrbd-20251008-213720-24811/junit.xml b/artifacts/m-cephrbd-20251008-213720-24811/junit.xml new file mode 100644 index 0000000000..22eea166c5 --- /dev/null +++ b/artifacts/m-cephrbd-20251008-213720-24811/junit.xml @@ -0,0 +1,878 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/08/25 22:03:55.808 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/08/25 22:04:00.545 (4.738s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/08/25 22:04:00.545 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/08/25 22:04:00.545 (0s) + + + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/08/25 22:04:00.549 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/08/25 22:04:02.578 (2.029s) > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/08/25 22:04:02.578 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/08/25 22:04:02.578 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/08/25 22:04:02.578 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/08/25 22:04:06.324 (3.746s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/08/25 22:04:06.324 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/08/25 22:04:06.324 (0s) + + + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/08/25 22:04:06.324 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/08/25 22:04:06.324 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/08/25 22:04:06.324 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:88 @ 10/08/25 22:04:06.324 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/08/25 22:04:15.04 (8.715s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/08/25 22:04:15.04 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/08/25 22:04:15.04 (0s) + + + [FAILED] Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-root-automatic-with-hotplug-standalone --namespace head-45bdb83d-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-root-automatic-with-hotplug-standalone\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/08/25 22:20:57.209 + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/08/25 22:04:15.04 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/08/25 22:04:15.04 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/08/25 22:04:15.04 STEP: VDs with consumers should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:99 @ 10/08/25 22:04:15.04 [FAILED] Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-root-automatic-with-hotplug-standalone --namespace head-45bdb83d-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-root-automatic-with-hotplug-standalone\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/08/25 22:20:57.209 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/08/25 22:20:57.21 (16m42.165s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/08/25 22:20:57.21 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-disk-attachment' error: exit status 1 stderr: E1008 22:20:57.257749 54438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.259554 54438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.261262 54438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.262951 54438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.264602 54438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.266242 54438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.267880 54438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.269530 54438 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/08/25 22:20:57.272 (62ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/08/25 22:20:57.272 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/08/25 22:20:57.272 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/08/25 22:20:57.272 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/08/25 22:20:57.272 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/08/25 22:20:57.272 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/08/25 22:20:57.272 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/08/25 22:20:57.272 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/08/25 22:20:57.272 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/08/25 22:20:57.272 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/08/25 22:20:57.272 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-restore-force Unexpected error: <*exec.ExitError | 0x14000a2a000>: exit status 1 { ProcessState: { pid: 54439, status: 256, rusage: { Utime: { Sec: 0, Usec: 23167, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7980, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44728320, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3535, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 42, Nvcsw: 21, Nivcsw: 445, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:61 @ 10/08/25 22:20:57.301 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/08/25 22:20:57.273 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-restore-force Unexpected error: <*exec.ExitError | 0x14000a2a000>: exit status 1 { ProcessState: { pid: 54439, status: 256, rusage: { Utime: { Sec: 0, Usec: 23167, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7980, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44728320, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3535, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 42, Nvcsw: 21, Nivcsw: 445, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:61 @ 10/08/25 22:20:57.301 < Exit [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/08/25 22:20:57.301 (29ms) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/08/25 22:20:57.301 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-restore-force' error: exit status 1 stderr: E1008 22:20:57.326381 54440 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.328121 54440 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.329829 54440 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.331511 54440 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.333128 54440 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.334765 54440 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.336391 54440 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.338105 54440 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? [PANICKED] Test Panicked In [AfterEach] at: /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/panic.go:262 @ 10/08/25 22:20:57.34 runtime error: invalid memory address or nil pointer dereference Full Stack Trace github.com/deckhouse/virtualization/tests/e2e.init.func21.3() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:76 +0xd4 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/08/25 22:20:57.34 (39ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/08/25 22:20:57.34 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:131 @ 10/08/25 22:20:57.34 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:177 @ 10/08/25 22:20:57.34 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:341 @ 10/08/25 22:20:57.34 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:367 @ 10/08/25 22:20:57.34 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-ipam Unexpected error: <*exec.ExitError | 0x14000a2a2c0>: exit status 1 { ProcessState: { pid: 54441, status: 256, rusage: { Utime: { Sec: 0, Usec: 20307, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7504, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44662784, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3533, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 39, Nvcsw: 11, Nivcsw: 437, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:56 @ 10/08/25 22:20:57.366 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/08/25 22:20:57.341 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-ipam Unexpected error: <*exec.ExitError | 0x14000a2a2c0>: exit status 1 { ProcessState: { pid: 54441, status: 256, rusage: { Utime: { Sec: 0, Usec: 20307, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7504, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44662784, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3533, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 39, Nvcsw: 11, Nivcsw: 437, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:56 @ 10/08/25 22:20:57.366 < Exit [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/08/25 22:20:57.366 (25ms) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/08/25 22:20:57.366 [PANICKED] Test Panicked In [AfterEach] at: /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/panic.go:262 @ 10/08/25 22:20:57.366 runtime error: invalid memory address or nil pointer dereference Full Stack Trace github.com/deckhouse/virtualization/tests/e2e.init.func8.5() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:158 +0x20 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/08/25 22:20:57.366 (0s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/08/25 22:20:57.366 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-complex-test Unexpected error: <*exec.ExitError | 0x140004d8000>: exit status 1 { ProcessState: { pid: 54442, status: 256, rusage: { Utime: { Sec: 0, Usec: 19697, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7683, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44695552, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3532, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 52, Nvcsw: 41, Nivcsw: 461, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:57 @ 10/08/25 22:20:57.39 + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:51 @ 10/08/25 22:20:57.366 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-complex-test Unexpected error: <*exec.ExitError | 0x140004d8000>: exit status 1 { ProcessState: { pid: 54442, status: 256, rusage: { Utime: { Sec: 0, Usec: 19697, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7683, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44695552, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3532, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 52, Nvcsw: 41, Nivcsw: 461, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:57 @ 10/08/25 22:20:57.39 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:51 @ 10/08/25 22:20:57.39 (24ms) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/08/25 22:20:57.39 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=complex-test' error: exit status 1 stderr: E1008 22:20:57.413596 54443 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.415306 54443 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.416934 54443 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.418598 54443 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.420351 54443 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.421975 54443 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.423616 54443 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.425269 54443 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/08/25 22:20:57.427 (37ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:82 @ 10/08/25 22:20:57.427 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:97 @ 10/08/25 22:20:57.427 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:108 @ 10/08/25 22:20:57.427 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:119 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:137 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:148 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:158 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:169 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:180 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:192 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:216 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:240 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:263 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:275 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:293 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:310 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:334 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:347 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:351 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:368 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:381 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:398 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:437 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:472 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:490 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:505 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:521 @ 10/08/25 22:20:57.428 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:538 @ 10/08/25 22:20:57.428 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-versions Unexpected error: <*exec.ExitError | 0x14000356280>: exit status 1 { ProcessState: { pid: 54444, status: 256, rusage: { Utime: { Sec: 0, Usec: 19627, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7569, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45383680, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3581, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 52, Nvcsw: 34, Nivcsw: 519, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:41 @ 10/08/25 22:20:57.452 + > Enter [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/08/25 22:20:57.428 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-versions Unexpected error: <*exec.ExitError | 0x14000356280>: exit status 1 { ProcessState: { pid: 54444, status: 256, rusage: { Utime: { Sec: 0, Usec: 19627, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7569, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45383680, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3581, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 52, Nvcsw: 34, Nivcsw: 519, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:41 @ 10/08/25 22:20:57.452 < Exit [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/08/25 22:20:57.452 (24ms) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/08/25 22:20:57.452 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-versions' error: exit status 1 stderr: E1008 22:20:57.476092 54445 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.477740 54445 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.479291 54445 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.480869 54445 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.482418 54445 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.483934 54445 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.485565 54445 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.487158 54445 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/08/25 22:20:57.489 (37ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/08/25 22:20:57.489 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/08/25 22:20:57.489 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/08/25 22:20:57.489 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/08/25 22:20:57.489 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/08/25 22:20:57.489 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vd-snapshots Unexpected error: <*exec.ExitError | 0x140008b3200>: exit status 1 { ProcessState: { pid: 54446, status: 256, rusage: { Utime: { Sec: 0, Usec: 19633, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7411, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45989888, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3611, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 48, Nvcsw: 23, Nivcsw: 448, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:64 @ 10/08/25 22:20:57.514 + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:54 @ 10/08/25 22:20:57.489 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vd-snapshots Unexpected error: <*exec.ExitError | 0x140008b3200>: exit status 1 { ProcessState: { pid: 54446, status: 256, rusage: { Utime: { Sec: 0, Usec: 19633, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7411, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45989888, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3611, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 48, Nvcsw: 23, Nivcsw: 448, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:64 @ 10/08/25 22:20:57.514 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:54 @ 10/08/25 22:20:57.514 (24ms) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/08/25 22:20:57.514 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vd-snapshots,id=head-45bdb83d' error: exit status 1 stderr: E1008 22:20:57.537054 54447 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.538783 54447 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.540556 54447 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.542262 54447 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.543949 54447 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.545487 54447 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.547171 54447 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.548768 54447 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/08/25 22:20:57.551 (37ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:95 @ 10/08/25 22:20:57.551 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:106 @ 10/08/25 22:20:57.551 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:117 @ 10/08/25 22:20:57.551 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:128 @ 10/08/25 22:20:57.551 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:139 @ 10/08/25 22:20:57.551 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:159 @ 10/08/25 22:20:57.551 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:180 @ 10/08/25 22:20:57.551 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:224 @ 10/08/25 22:20:57.551 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:285 @ 10/08/25 22:20:57.551 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:331 @ 10/08/25 22:20:57.551 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:357 @ 10/08/25 22:20:57.551 + + + [FAILED] Unexpected error: <*exec.ExitError | 0x140008b3c20>: exit status 1 { ProcessState: { pid: 54448, status: 256, rusage: { Utime: { Sec: 0, Usec: 22271, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 8761, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45268992, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3567, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 45, Msgrcv: 45, Nsignals: 50, Nvcsw: 55, Nivcsw: 576, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:65 @ 10/08/25 22:20:57.586 + > Enter [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/08/25 22:20:57.551 [FAILED] Unexpected error: <*exec.ExitError | 0x140008b3c20>: exit status 1 { ProcessState: { pid: 54448, status: 256, rusage: { Utime: { Sec: 0, Usec: 22271, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 8761, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45268992, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3567, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 45, Msgrcv: 45, Nsignals: 50, Nvcsw: 55, Nivcsw: 576, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:65 @ 10/08/25 22:20:57.586 < Exit [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/08/25 22:20:57.586 (35ms) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/08/25 22:20:57.586 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=image-hotplug' error: exit status 1 stderr: E1008 22:20:57.609959 54449 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.611559 54449 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.613350 54449 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.614900 54449 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.616460 54449 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.618032 54449 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.619615 54449 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:57.621207 54449 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/08/25 22:20:57.623 (37ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:112 @ 10/08/25 22:20:57.623 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:147 @ 10/08/25 22:20:57.623 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:190 @ 10/08/25 22:20:57.623 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:196 @ 10/08/25 22:20:57.623 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:204 @ 10/08/25 22:20:57.623 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:224 @ 10/08/25 22:20:57.623 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:236 @ 10/08/25 22:20:57.623 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:258 @ 10/08/25 22:20:57.623 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:284 @ 10/08/25 22:20:57.623 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:294 @ 10/08/25 22:20:57.623 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:309 @ 10/08/25 22:20:57.623 + + + + > Enter [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/08/25 22:20:57.623 [SKIPPED] Module SDN is disabled. Skipping all tests for module SDN. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:50 @ 10/08/25 22:20:58.229 < Exit [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/08/25 22:20:58.229 (606ms) > Enter [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/08/25 22:20:58.229 < Exit [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/08/25 22:20:58.229 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/08/25 22:20:58.229 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/08/25 22:20:58.23 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:111 @ 10/08/25 22:20:58.23 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:125 @ 10/08/25 22:20:58.23 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:141 @ 10/08/25 22:20:58.23 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:156 @ 10/08/25 22:20:58.23 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:169 @ 10/08/25 22:20:58.23 + + + [FAILED] error: error validating "/tmp/testdata/importer-network-policy/project": error validating data: failed to download openapi: Get "http://localhost:8080/openapi/v2?timeout=32s": dial tcp [::1]:8080: connect: connection refused; if you choose to ignore these errors, turn validation off with --validate=false Expected <bool>: false to equal <bool>: true In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:69 @ 10/08/25 22:20:58.294 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/08/25 22:20:58.231 < Exit [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/08/25 22:20:58.232 (1ms) > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/08/25 22:20:58.232 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/08/25 22:20:58.232 (0s) > Enter [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/08/25 22:20:58.232 [FAILED] error: error validating "/tmp/testdata/importer-network-policy/project": error validating data: failed to download openapi: Get "http://localhost:8080/openapi/v2?timeout=32s": dial tcp [::1]:8080: connect: connection refused; if you choose to ignore these errors, turn validation off with --validate=false Expected <bool>: false to equal <bool>: true In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:69 @ 10/08/25 22:20:58.294 < Exit [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/08/25 22:20:58.294 (62ms) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/08/25 22:20:58.294 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=importer-network-policy' error: exit status 1 stderr: E1008 22:20:58.320838 54452 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.322624 54452 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.324276 54452 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.325888 54452 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.327419 54452 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.328995 54452 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.330641 54452 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.332262 54452 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/08/25 22:20:58.334 (40ms) > Enter [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/08/25 22:20:58.334 STEP: Delete manifests - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:50 @ 10/08/25 22:20:58.334 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/08/25 22:20:58.334 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/08/25 22:20:58.374 (39ms) [FAILED] cannot delete test case resources kustomizationDir: /tmp/testdata/importer-network-policy cmd: kubectl delete --kustomize /tmp/testdata/importer-network-policy --ignore-not-found stderr: E1008 22:20:58.364968 54453 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.366727 54453 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.368261 54453 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.369885 54453 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.371534 54453 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" unable to recognize "/tmp/testdata/importer-network-policy": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/importer-network-policy": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/importer-network-policy": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/importer-network-policy": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/importer-network-policy": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused Unexpected error: <*exec.ExitError | 0x14000b002a0>: exit status 1 { ProcessState: { pid: 54453, status: 256, rusage: { Utime: { Sec: 0, Usec: 29677, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 10772, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 48136192, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3745, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 33, Msgrcv: 33, Nsignals: 75, Nvcsw: 140, Nivcsw: 678, }, }, Stderr: nil, } occurred In [AfterAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:617 @ 10/08/25 22:20:58.374 < Exit [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/08/25 22:20:58.374 (40ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/08/25 22:20:58.374 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/08/25 22:20:58.374 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/08/25 22:20:58.374 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/08/25 22:20:58.374 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/08/25 22:20:58.374 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-label-annotation Unexpected error: <*exec.ExitError | 0x140004d8000>: exit status 1 { ProcessState: { pid: 54454, status: 256, rusage: { Utime: { Sec: 0, Usec: 19448, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7032, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44662784, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3532, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 40, Nvcsw: 16, Nivcsw: 434, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:48 @ 10/08/25 22:20:58.398 + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/08/25 22:20:58.374 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-label-annotation Unexpected error: <*exec.ExitError | 0x140004d8000>: exit status 1 { ProcessState: { pid: 54454, status: 256, rusage: { Utime: { Sec: 0, Usec: 19448, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7032, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44662784, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3532, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 40, Nvcsw: 16, Nivcsw: 434, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:48 @ 10/08/25 22:20:58.398 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/08/25 22:20:58.398 (25ms) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/08/25 22:20:58.398 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-label-annotation' error: exit status 1 stderr: E1008 22:20:58.421630 54455 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.423410 54455 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.425020 54455 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.426641 54455 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.428308 54455 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.429933 54455 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.431441 54455 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.433094 54455 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/08/25 22:20:58.435 (37ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/08/25 22:20:58.435 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/08/25 22:20:58.435 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/08/25 22:20:58.435 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/08/25 22:20:58.435 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/08/25 22:20:58.435 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/08/25 22:20:58.435 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/08/25 22:20:58.435 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/08/25 22:20:58.435 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/08/25 22:20:58.435 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/08/25 22:20:58.435 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/08/25 22:20:58.435 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/08/25 22:20:58.435 + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.435 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:20:58.436 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:20:58.436 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:20:58.436 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:20:58.436 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:20:58.436 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:20:58.436 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:20:58.436 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:20:58.436 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:20:58.436 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:20:58.436 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:20:58.436 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:20:58.436 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:20:58.436 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:20:58.437 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:20:58.437 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:20:58.437 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:20:58.437 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:20:58.437 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:20:58.437 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:20:58.437 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:20:58.437 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:20:58.437 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:20:58.437 (0s) + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-migration Unexpected error: <*exec.ExitError | 0x140004d8460>: exit status 1 { ProcessState: { pid: 54456, status: 256, rusage: { Utime: { Sec: 0, Usec: 19479, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7490, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45432832, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3582, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 45, Nvcsw: 36, Nivcsw: 441, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:42 @ 10/08/25 22:20:58.461 + > Enter [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/08/25 22:20:58.437 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-migration Unexpected error: <*exec.ExitError | 0x140004d8460>: exit status 1 { ProcessState: { pid: 54456, status: 256, rusage: { Utime: { Sec: 0, Usec: 19479, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7490, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45432832, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3582, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 45, Nvcsw: 36, Nivcsw: 441, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:42 @ 10/08/25 22:20:58.461 < Exit [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/08/25 22:20:58.461 (24ms) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/08/25 22:20:58.461 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-migration' error: exit status 1 stderr: E1008 22:20:58.484725 54457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.486504 54457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.488201 54457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.489849 54457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.491485 54457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.493152 54457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.494814 54457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.496422 54457 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/08/25 22:20:58.498 (37ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/08/25 22:20:58.499 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:86 @ 10/08/25 22:20:58.499 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:100 @ 10/08/25 22:20:58.499 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:116 @ 10/08/25 22:20:58.499 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:131 @ 10/08/25 22:20:58.499 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-configuration Unexpected error: <*exec.ExitError | 0x140008b2100>: exit status 1 { ProcessState: { pid: 54458, status: 256, rusage: { Utime: { Sec: 0, Usec: 19874, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 6765, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45088768, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3561, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 38, Nvcsw: 17, Nivcsw: 428, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:55 @ 10/08/25 22:20:58.523 + > Enter [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/08/25 22:20:58.499 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-configuration Unexpected error: <*exec.ExitError | 0x140008b2100>: exit status 1 { ProcessState: { pid: 54458, status: 256, rusage: { Utime: { Sec: 0, Usec: 19874, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 6765, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45088768, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3561, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 38, Nvcsw: 17, Nivcsw: 428, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:55 @ 10/08/25 22:20:58.523 < Exit [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/08/25 22:20:58.523 (25ms) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/08/25 22:20:58.523 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-configuration' error: exit status 1 stderr: E1008 22:20:58.546600 54459 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.548293 54459 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.550002 54459 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.551605 54459 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.553579 54459 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.555182 54459 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.556851 54459 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.558517 54459 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/08/25 22:20:58.56 (37ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/08/25 22:20:58.56 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/08/25 22:20:58.56 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:109 @ 10/08/25 22:20:58.56 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:123 @ 10/08/25 22:20:58.56 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:147 @ 10/08/25 22:20:58.56 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:161 @ 10/08/25 22:20:58.56 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:183 @ 10/08/25 22:20:58.561 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:202 @ 10/08/25 22:20:58.561 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:226 @ 10/08/25 22:20:58.561 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:240 @ 10/08/25 22:20:58.561 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:250 @ 10/08/25 22:20:58.561 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:265 @ 10/08/25 22:20:58.561 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-images-creation Unexpected error: <*exec.ExitError | 0x140008b2aa0>: exit status 1 { ProcessState: { pid: 54460, status: 256, rusage: { Utime: { Sec: 0, Usec: 19657, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7068, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44777472, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3538, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 33, Nvcsw: 14, Nivcsw: 430, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:48 @ 10/08/25 22:20:58.585 + > Enter [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:38 @ 10/08/25 22:20:58.561 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-images-creation Unexpected error: <*exec.ExitError | 0x140008b2aa0>: exit status 1 { ProcessState: { pid: 54460, status: 256, rusage: { Utime: { Sec: 0, Usec: 19657, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7068, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44777472, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3538, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 33, Nvcsw: 14, Nivcsw: 430, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:48 @ 10/08/25 22:20:58.585 < Exit [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:38 @ 10/08/25 22:20:58.585 (25ms) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/08/25 22:20:58.585 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=images-creation' error: exit status 1 stderr: E1008 22:20:58.608877 54461 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.610674 54461 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.612324 54461 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.613933 54461 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.615538 54461 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.617134 54461 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.618783 54461 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.620426 54461 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/08/25 22:20:58.622 (37ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:87 @ 10/08/25 22:20:58.623 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:96 @ 10/08/25 22:20:58.623 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:107 @ 10/08/25 22:20:58.623 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:116 @ 10/08/25 22:20:58.623 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:127 @ 10/08/25 22:20:58.623 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-restore-safe Unexpected error: <*exec.ExitError | 0x14000a2bb00>: exit status 1 { ProcessState: { pid: 54462, status: 256, rusage: { Utime: { Sec: 0, Usec: 19576, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7135, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44974080, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3555, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 41, Nvcsw: 20, Nivcsw: 460, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:58 @ 10/08/25 22:20:58.647 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/08/25 22:20:58.623 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-restore-safe Unexpected error: <*exec.ExitError | 0x14000a2bb00>: exit status 1 { ProcessState: { pid: 54462, status: 256, rusage: { Utime: { Sec: 0, Usec: 19576, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7135, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44974080, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3555, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 41, Nvcsw: 20, Nivcsw: 460, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:58 @ 10/08/25 22:20:58.647 < Exit [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/08/25 22:20:58.647 (25ms) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/08/25 22:20:58.647 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-restore-safe' error: exit status 1 stderr: E1008 22:20:58.670835 54463 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.672549 54463 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.674207 54463 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.675821 54463 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.677517 54463 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.679124 54463 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.680762 54463 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.682412 54463 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? [PANICKED] Test Panicked In [AfterEach] at: /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/panic.go:262 @ 10/08/25 22:20:58.684 runtime error: invalid memory address or nil pointer dereference Full Stack Trace github.com/deckhouse/virtualization/tests/e2e.init.func22.3() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:70 +0xd4 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/08/25 22:20:58.684 (37ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/08/25 22:20:58.684 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:115 @ 10/08/25 22:20:58.684 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:158 @ 10/08/25 22:20:58.684 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:348 @ 10/08/25 22:20:58.684 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:374 @ 10/08/25 22:20:58.684 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-sizing-policy Unexpected error: <*exec.ExitError | 0x14000a2bd80>: exit status 1 { ProcessState: { pid: 54464, status: 256, rusage: { Utime: { Sec: 0, Usec: 19144, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7394, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44449792, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3518, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 46, Nvcsw: 27, Nivcsw: 445, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:64 @ 10/08/25 22:20:58.708 + > Enter [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/08/25 22:20:58.684 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-sizing-policy Unexpected error: <*exec.ExitError | 0x14000a2bd80>: exit status 1 { ProcessState: { pid: 54464, status: 256, rusage: { Utime: { Sec: 0, Usec: 19144, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7394, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44449792, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3518, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 46, Nvcsw: 27, Nivcsw: 445, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:64 @ 10/08/25 22:20:58.708 < Exit [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/08/25 22:20:58.708 (24ms) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/08/25 22:20:58.708 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=sizing-policy' error: exit status 1 stderr: E1008 22:20:58.733848 54465 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.735552 54465 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.737374 54465 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.739031 54465 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.740697 54465 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.742172 54465 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.743867 54465 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.745516 54465 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/08/25 22:20:58.747 (39ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/08/25 22:20:58.748 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/08/25 22:20:58.748 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/08/25 22:20:58.748 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:126 @ 10/08/25 22:20:58.748 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:140 @ 10/08/25 22:20:58.748 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:152 @ 10/08/25 22:20:58.748 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:157 @ 10/08/25 22:20:58.748 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:163 @ 10/08/25 22:20:58.748 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:176 @ 10/08/25 22:20:58.748 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:181 @ 10/08/25 22:20:58.748 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:187 @ 10/08/25 22:20:58.748 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:202 @ 10/08/25 22:20:58.748 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:216 @ 10/08/25 22:20:58.748 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:240 @ 10/08/25 22:20:58.748 + + + + > Enter [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/08/25 22:20:58.748 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:51 @ 10/08/25 22:20:58.748 < Exit [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/08/25 22:20:58.748 (0s) > Enter [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/08/25 22:20:58.748 < Exit [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/08/25 22:20:58.749 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:82 @ 10/08/25 22:20:58.749 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:94 @ 10/08/25 22:20:58.749 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:106 @ 10/08/25 22:20:58.749 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:116 @ 10/08/25 22:20:58.749 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:132 @ 10/08/25 22:20:58.749 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:151 @ 10/08/25 22:20:58.749 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:157 @ 10/08/25 22:20:58.749 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:189 @ 10/08/25 22:20:58.749 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:216 @ 10/08/25 22:20:58.749 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:252 @ 10/08/25 22:20:58.749 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-evacuation Unexpected error: <*exec.ExitError | 0x140008b2100>: exit status 1 { ProcessState: { pid: 54467, status: 256, rusage: { Utime: { Sec: 0, Usec: 19953, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7486, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44204032, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3505, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 40, Nvcsw: 22, Nivcsw: 450, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:48 @ 10/08/25 22:20:58.774 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/08/25 22:20:58.749 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-evacuation Unexpected error: <*exec.ExitError | 0x140008b2100>: exit status 1 { ProcessState: { pid: 54467, status: 256, rusage: { Utime: { Sec: 0, Usec: 19953, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7486, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44204032, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3505, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 40, Nvcsw: 22, Nivcsw: 450, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:48 @ 10/08/25 22:20:58.774 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/08/25 22:20:58.774 (25ms) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/08/25 22:20:58.774 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-evacuation' error: exit status 1 stderr: E1008 22:20:58.797513 54468 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.799297 54468 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.801051 54468 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.802735 54468 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.804379 54468 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.806054 54468 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.807695 54468 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.809349 54468 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/08/25 22:20:58.811 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/08/25 22:20:58.883 (71ms) [FAILED] cannot delete test case resources kustomizationDir: /tmp/testdata/vm-evacuation cmd: kubectl delete --kustomize /tmp/testdata/vm-evacuation --ignore-not-found stderr: E1008 22:20:58.865449 54469 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.867108 54469 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.868690 54469 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.870304 54469 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.871954 54469 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.873443 54469 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.875045 54469 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.876725 54469 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.878407 54469 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.880065 54469 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused Unexpected error: <*exec.ExitError | 0x14000b00000>: exit status 1 { ProcessState: { pid: 54469, status: 256, rusage: { Utime: { Sec: 0, Usec: 79783, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 14979, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 79282176, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5647, Majflt: 4, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 63, Msgrcv: 63, Nsignals: 137, Nvcsw: 136, Nivcsw: 1045, }, }, Stderr: nil, } occurred In [AfterEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:617 @ 10/08/25 22:20:58.883 < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/08/25 22:20:58.883 (109ms) + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-connectivity Unexpected error: <*exec.ExitError | 0x140002b5ea0>: exit status 1 { ProcessState: { pid: 54470, status: 256, rusage: { Utime: { Sec: 0, Usec: 19611, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7528, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44924928, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3545, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 42, Nvcsw: 52, Nivcsw: 436, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:63 @ 10/08/25 22:20:58.908 + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/08/25 22:20:58.883 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-connectivity Unexpected error: <*exec.ExitError | 0x140002b5ea0>: exit status 1 { ProcessState: { pid: 54470, status: 256, rusage: { Utime: { Sec: 0, Usec: 19611, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7528, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44924928, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3545, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 42, Nvcsw: 52, Nivcsw: 436, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:63 @ 10/08/25 22:20:58.908 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/08/25 22:20:58.908 (25ms) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/08/25 22:20:58.908 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-connectivity' error: exit status 1 stderr: E1008 22:20:58.931151 54471 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.932865 54471 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.934515 54471 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.936145 54471 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.937913 54471 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.939630 54471 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.941316 54471 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:58.942980 54471 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/08/25 22:20:58.945 (37ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/08/25 22:20:58.945 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/08/25 22:20:58.945 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:118 @ 10/08/25 22:20:58.945 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:129 @ 10/08/25 22:20:58.945 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:142 @ 10/08/25 22:20:58.945 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:166 @ 10/08/25 22:20:58.945 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:174 @ 10/08/25 22:20:58.945 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:179 @ 10/08/25 22:20:58.945 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:187 @ 10/08/25 22:20:58.945 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:198 @ 10/08/25 22:20:58.945 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:209 @ 10/08/25 22:20:58.945 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:222 @ 10/08/25 22:20:58.945 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:228 @ 10/08/25 22:20:58.945 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:240 @ 10/08/25 22:20:58.945 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:250 @ 10/08/25 22:20:58.945 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:258 @ 10/08/25 22:20:58.945 + + + + > Enter [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/08/25 22:20:58.945 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:41 @ 10/08/25 22:20:58.945 < Exit [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/08/25 22:20:58.945 (0s) > Enter [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/08/25 22:20:58.945 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/08/25 22:20:58.945 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/08/25 22:20:59.015 (70ms) [FAILED] cannot delete test case resources kustomizationDir: /tmp/testdata/vm-migration-cancel cmd: kubectl delete --kustomize /tmp/testdata/vm-migration-cancel --ignore-not-found stderr: E1008 22:20:58.999091 54472 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.000619 54472 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.002052 54472 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.003469 54472 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.005008 54472 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.006614 54472 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.008141 54472 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.009641 54472 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.011249 54472 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.012837 54472 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused Unexpected error: <*exec.ExitError | 0x14000357140>: exit status 1 { ProcessState: { pid: 54472, status: 256, rusage: { Utime: { Sec: 0, Usec: 78603, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 15329, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 79265792, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5648, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 63, Msgrcv: 63, Nsignals: 135, Nvcsw: 150, Nivcsw: 1087, }, }, Stderr: nil, } occurred In [AfterEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:617 @ 10/08/25 22:20:59.015 < Exit [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/08/25 22:20:59.015 (70ms) + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-affinity-toleration Unexpected error: <*exec.ExitError | 0x1400059fba0>: exit status 1 { ProcessState: { pid: 54474, status: 256, rusage: { Utime: { Sec: 0, Usec: 20920, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7589, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44531712, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3526, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 47, Nvcsw: 38, Nivcsw: 468, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:69 @ 10/08/25 22:20:59.04 + > Enter [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/08/25 22:20:59.015 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-affinity-toleration Unexpected error: <*exec.ExitError | 0x1400059fba0>: exit status 1 { ProcessState: { pid: 54474, status: 256, rusage: { Utime: { Sec: 0, Usec: 20920, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 7589, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44531712, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3526, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 47, Nvcsw: 38, Nivcsw: 468, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:69 @ 10/08/25 22:20:59.04 < Exit [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/08/25 22:20:59.04 (25ms) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/08/25 22:20:59.04 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=affinity-toleration' error: exit status 1 stderr: E1008 22:20:59.063605 54475 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.065337 54475 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.067003 54475 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.068594 54475 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.070217 54475 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.071766 54475 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.073537 54475 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.075191 54475 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/08/25 22:20:59.077 (37ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/08/25 22:20:59.077 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:120 @ 10/08/25 22:20:59.077 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:283 @ 10/08/25 22:20:59.077 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:372 @ 10/08/25 22:20:59.077 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:467 @ 10/08/25 22:20:59.077 + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x14000afeed0>: the container "virtualization-controller" was not found: virtualization-controller-5c7ffcff8c-9qksd the container "virtualization-controller" was restarted: virtualization-controller-c9bc9b75c-9tclq the container "virtualization-controller" was restarted: virtualization-controller-c9bc9b75c-hr9xw { errs: [ <*errors.joinError | 0x14000afeeb8>{ errs: [ <*errors.joinError | 0x14000afeea0>{ errs: [ <*errors.errorString | 0x140002c2220>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-5c7ffcff8c-9qksd", }, ], }, <*errors.errorString | 0x140002c2260>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-c9bc9b75c-9tclq", }, ], }, <*errors.errorString | 0x140002c2300>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-c9bc9b75c-hr9xw", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/08/25 22:20:59.332 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/08/25 22:20:59.077 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/08/25 22:20:59.077 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/08/25 22:20:59.077 [FAILED] Expected success, but got an error: <*errors.joinError | 0x14000afeed0>: the container "virtualization-controller" was not found: virtualization-controller-5c7ffcff8c-9qksd the container "virtualization-controller" was restarted: virtualization-controller-c9bc9b75c-9tclq the container "virtualization-controller" was restarted: virtualization-controller-c9bc9b75c-hr9xw { errs: [ <*errors.joinError | 0x14000afeeb8>{ errs: [ <*errors.joinError | 0x14000afeea0>{ errs: [ <*errors.errorString | 0x140002c2220>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-5c7ffcff8c-9qksd", }, ], }, <*errors.errorString | 0x140002c2260>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-c9bc9b75c-9tclq", }, ], }, <*errors.errorString | 0x140002c2300>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-c9bc9b75c-hr9xw", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/08/25 22:20:59.332 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/08/25 22:20:59.332 (255ms) + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x14000aff1a0>: cmd: kubectl delete clustervirtualimages.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found stderr: E1008 22:20:59.492932 54476 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.494771 54476 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.496475 54476 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.498259 54476 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.499992 54476 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.501686 54476 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.503351 54476 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? cmd: kubectl delete virtualmachineclasses.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found stderr: E1008 22:20:59.535013 54477 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.536766 54477 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.538452 54477 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.540150 54477 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.541823 54477 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.543394 54477 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.545068 54477 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? cmd: kubectl delete virtualmachineipaddressleases.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found stderr: E1008 22:20:59.573029 54478 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.574597 54478 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.576137 54478 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.577680 54478 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.579240 54478 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.580782 54478 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.582313 54478 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? { errs: [ <*errors.joinError | 0x14000aff0c8>{ errs: [ <*errors.joinError | 0x140004262b8>{ errs: [ <*errors.errorString | 0x140003eeff0>{ s: "cmd: kubectl delete clustervirtualimages.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found\nstderr: E1008 22:20:59.492932 54476 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.494771 54476 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.496475 54476 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.498259 54476 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.499992 54476 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.501686 54476 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.503351 54476 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nThe connection to the server localhost:8080 was refused - did you specify the right host or port?\n", }, ], }, <*errors.errorString | 0x140002c38d0>{ s: "cmd: kubectl delete virtualmachineclasses.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found\nstderr: E1008 22:20:59.535013 54477 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.536766 54477 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.538452 54477 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.540150 54477 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.541823 54477 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.543394 54477 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.545068 54477 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nThe connection to the server localhost:8080 was refused - did you specify the right host or port?\n", }, ... Gomega truncated this representation as it exceeds 'format.MaxLength'. Consider having the object provide a custom 'GomegaStringer' representation or adjust the parameters in Gomega's 'format' package. Learn more here: https://onsi.github.io/gomega/#adjusting-output In [DeferCleanup (Suite)] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:177 @ 10/08/25 22:21:01.758 + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/08/25 22:20:59.332 [FAILED] Expected success, but got an error: <*errors.joinError | 0x14000aff1a0>: cmd: kubectl delete clustervirtualimages.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found stderr: E1008 22:20:59.492932 54476 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.494771 54476 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.496475 54476 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.498259 54476 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.499992 54476 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.501686 54476 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.503351 54476 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? cmd: kubectl delete virtualmachineclasses.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found stderr: E1008 22:20:59.535013 54477 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.536766 54477 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.538452 54477 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.540150 54477 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.541823 54477 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.543394 54477 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.545068 54477 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? cmd: kubectl delete virtualmachineipaddressleases.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found stderr: E1008 22:20:59.573029 54478 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.574597 54478 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.576137 54478 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.577680 54478 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.579240 54478 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.580782 54478 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:20:59.582313 54478 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? { errs: [ <*errors.joinError | 0x14000aff0c8>{ errs: [ <*errors.joinError | 0x140004262b8>{ errs: [ <*errors.errorString | 0x140003eeff0>{ s: "cmd: kubectl delete clustervirtualimages.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found\nstderr: E1008 22:20:59.492932 54476 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.494771 54476 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.496475 54476 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.498259 54476 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.499992 54476 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.501686 54476 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.503351 54476 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nThe connection to the server localhost:8080 was refused - did you specify the right host or port?\n", }, ], }, <*errors.errorString | 0x140002c38d0>{ s: "cmd: kubectl delete virtualmachineclasses.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found\nstderr: E1008 22:20:59.535013 54477 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.536766 54477 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.538452 54477 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.540150 54477 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.541823 54477 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.543394 54477 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:20:59.545068 54477 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nThe connection to the server localhost:8080 was refused - did you specify the right host or port?\n", }, ... Gomega truncated this representation as it exceeds 'format.MaxLength'. Consider having the object provide a custom 'GomegaStringer' representation or adjust the parameters in Gomega's 'format' package. Learn more here: https://onsi.github.io/gomega/#adjusting-output In [DeferCleanup (Suite)] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:177 @ 10/08/25 22:21:01.758 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/08/25 22:21:01.759 (2.426s) + + + \ No newline at end of file diff --git a/artifacts/m-sds-20251008-213720-16317/junit.xml b/artifacts/m-sds-20251008-213720-16317/junit.xml new file mode 100644 index 0000000000..be3ee27c35 --- /dev/null +++ b/artifacts/m-sds-20251008-213720-16317/junit.xml @@ -0,0 +1,871 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/08/25 22:02:38.254 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/08/25 22:02:42.879 (4.625s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/08/25 22:02:42.879 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/08/25 22:02:42.879 (0s) + + + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:51 @ 10/08/25 22:02:42.881 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:51 @ 10/08/25 22:02:44.833 (1.953s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:61 @ 10/08/25 22:02:44.834 �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Failed to update resource", "err": "error updating status subresource: Put \"http://127.0.0.1:23915/apis/virtualization.deckhouse.io/v1alpha2/clustervirtualimages/head-45bdb83d-cvi-from-cvi-ubu-http/status\": context canceled", "controller": "cvi-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-cvi-from-cvi-ubu-http", "namespace": "", "reconcileID": "d768e99d-88b1-4646-ad8d-bdcac2dd909b", "time": "2025-10-08T19:03:06Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "error updating status subresource: Put \"http://127.0.0.1:23915/apis/virtualization.deckhouse.io/v1alpha2/clustervirtualimages/head-45bdb83d-cvi-from-cvi-ubu-http/status\": context canceled", "controller": "cvi-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-cvi-from-cvi-ubu-http", "namespace": "", "reconcileID": "d768e99d-88b1-4646-ad8d-bdcac2dd909b", "time": "2025-10-08T19:03:06Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "error updating status subresource: Put \"http://127.0.0.1:23915/apis/virtualization.deckhouse.io/v1alpha2/clustervirtualimages/head-45bdb83d-cvi-from-cvi-ubu-http/status\": context canceled", "controller": "cvi-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-cvi-from-cvi-ubu-http", "namespace": "", "reconcileID": "d768e99d-88b1-4646-ad8d-bdcac2dd909b", "time": "2025-10-08T19:03:06Z" }�[0m < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:61 @ 10/08/25 22:03:13.333 (28.499s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/08/25 22:03:13.333 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/08/25 22:03:13.333 (0s) + + + > Enter [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:82 @ 10/08/25 22:03:13.334 < Exit [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:82 @ 10/08/25 22:03:14.552 (1.218s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/08/25 22:03:14.552 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/08/25 22:03:14.552 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:97 @ 10/08/25 22:03:14.553 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:98 @ 10/08/25 22:03:14.554 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:97 @ 10/08/25 22:03:15.887 (1.334s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/08/25 22:03:15.887 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/08/25 22:03:15.887 (0s) + + + > Enter [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:108 @ 10/08/25 22:03:15.887 STEP: CVIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:109 @ 10/08/25 22:03:15.888 < Exit [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:108 @ 10/08/25 22:03:16.983 (1.096s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/08/25 22:03:16.983 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/08/25 22:03:16.984 (0s) + + + > Enter [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:119 @ 10/08/25 22:03:16.984 STEP: VMClasses should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:120 @ 10/08/25 22:03:16.984 < Exit [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:119 @ 10/08/25 22:03:18.354 (1.37s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/08/25 22:03:18.354 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/08/25 22:03:18.354 (0s) + + + [FAILED] Timed out after 343.852s. Expected success, but got an error: <*errors.errorString | 0x1400006d7e0>: cannot patch VMIP "head-45bdb83d-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-45bdb83d-vm-custom-ip { s: "cannot patch VMIP \"head-45bdb83d-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-45bdb83d-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/08/25 22:09:02.209 + > Enter [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 @ 10/08/25 22:03:18.354 [FAILED] Timed out after 343.852s. Expected success, but got an error: <*errors.errorString | 0x1400006d7e0>: cannot patch VMIP "head-45bdb83d-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-45bdb83d-vm-custom-ip { s: "cannot patch VMIP \"head-45bdb83d-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-45bdb83d-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/08/25 22:09:02.209 < Exit [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 @ 10/08/25 22:09:02.209 (5m43.853s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/08/25 22:09:02.209 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/08/25 22:09:06.97 (4.761s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:137 @ 10/08/25 22:09:06.97 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:148 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:158 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:169 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:180 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:192 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:216 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:240 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:263 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:275 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:293 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:310 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:334 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:347 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:351 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:368 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:381 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:398 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:437 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:472 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:490 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:505 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:521 @ 10/08/25 22:09:06.971 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:538 @ 10/08/25 22:09:06.971 + + + > Enter [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/08/25 22:09:06.971 < Exit [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/08/25 22:09:09.587 (2.616s) > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/08/25 22:09:09.587 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/08/25 22:09:09.587 (0s) > Enter [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/08/25 22:09:09.587 STEP: Create a vmip automatically and check its binding with a lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:81 @ 10/08/25 22:09:09.587 STEP: Remove label from the lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:87 @ 10/08/25 22:09:09.936 STEP: Wait for the label to be restored by the controller - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:96 @ 10/08/25 22:09:10.051 < Exit [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/08/25 22:09:10.279 (692ms) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/08/25 22:09:10.279 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/08/25 22:09:10.279 (0s) + + + > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/08/25 22:09:10.279 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/08/25 22:09:10.279 (0s) > Enter [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/08/25 22:09:10.279 STEP: Create an intermediate vmip automatically to allocate a new ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:109 @ 10/08/25 22:09:10.279 STEP: Delete the intermediate vmip automatically and check that the lease is released - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:115 @ 10/08/25 22:09:10.638 STEP: Reuse the released lease with a static vmip - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:124 @ 10/08/25 22:09:10.871 STEP: Delete the static vmip and lease, then create another static vmip with this ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:132 @ 10/08/25 22:09:11.336 < Exit [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/08/25 22:09:11.984 (1.705s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/08/25 22:09:11.984 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/08/25 22:09:11.984 (0s) + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:66 @ 10/08/25 22:09:13.519 + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:54 @ 10/08/25 22:09:11.985 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:66 @ 10/08/25 22:09:13.519 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:54 @ 10/08/25 22:09:13.519 (1.534s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/08/25 22:09:13.519 < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/08/25 22:09:18.497 (4.978s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:95 @ 10/08/25 22:09:18.497 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:106 @ 10/08/25 22:09:18.497 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:117 @ 10/08/25 22:09:18.497 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:128 @ 10/08/25 22:09:18.498 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:139 @ 10/08/25 22:09:18.498 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:159 @ 10/08/25 22:09:18.498 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:180 @ 10/08/25 22:09:18.498 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:224 @ 10/08/25 22:09:18.498 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:285 @ 10/08/25 22:09:18.498 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:331 @ 10/08/25 22:09:18.498 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:357 @ 10/08/25 22:09:18.498 + + + > Enter [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/08/25 22:09:18.498 < Exit [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/08/25 22:09:20.023 (1.525s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/08/25 22:09:20.023 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/08/25 22:09:24.657 (4.634s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/08/25 22:09:24.657 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/08/25 22:09:24.657 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/08/25 22:09:24.657 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:89 @ 10/08/25 22:09:24.657 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/08/25 22:09:28.808 (4.151s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/08/25 22:09:28.808 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/08/25 22:09:28.808 (0s) + + + [FAILED] Expected <[]string | len:2, cap:4>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-root-manual-conf --namespace head-45bdb83d-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-root-manual-conf\n", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-root-automatic-conf --namespace head-45bdb83d-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-root-automatic-conf\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:100 @ 10/08/25 22:26:10.456 + > Enter [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/08/25 22:09:28.809 [FAILED] Expected <[]string | len:2, cap:4>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-root-manual-conf --namespace head-45bdb83d-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-root-manual-conf\n", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-root-automatic-conf --namespace head-45bdb83d-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-root-automatic-conf\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:100 @ 10/08/25 22:26:10.456 < Exit [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/08/25 22:26:10.456 (16m41.64s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/08/25 22:26:10.456 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-configuration' error: exit status 1 stderr: E1008 22:26:10.492406 54913 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.493923 54913 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.495403 54913 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.496903 54913 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.498415 54913 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.499859 54913 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.501282 54913 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.502750 54913 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/08/25 22:26:10.505 (49ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:109 @ 10/08/25 22:26:10.505 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:123 @ 10/08/25 22:26:10.505 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:147 @ 10/08/25 22:26:10.505 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:161 @ 10/08/25 22:26:10.505 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:183 @ 10/08/25 22:26:10.505 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:202 @ 10/08/25 22:26:10.505 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:226 @ 10/08/25 22:26:10.505 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:240 @ 10/08/25 22:26:10.505 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:250 @ 10/08/25 22:26:10.505 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:265 @ 10/08/25 22:26:10.505 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-restore-force Unexpected error: <*exec.ExitError | 0x140007165e0>: exit status 1 { ProcessState: { pid: 54914, status: 256, rusage: { Utime: { Sec: 0, Usec: 23659, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 8718, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44449792, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3521, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 54, Nvcsw: 48, Nivcsw: 443, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:61 @ 10/08/25 22:26:10.536 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/08/25 22:26:10.505 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-restore-force Unexpected error: <*exec.ExitError | 0x140007165e0>: exit status 1 { ProcessState: { pid: 54914, status: 256, rusage: { Utime: { Sec: 0, Usec: 23659, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 8718, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44449792, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3521, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 54, Nvcsw: 48, Nivcsw: 443, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:61 @ 10/08/25 22:26:10.536 < Exit [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/08/25 22:26:10.536 (31ms) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/08/25 22:26:10.536 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-restore-force' error: exit status 1 stderr: E1008 22:26:10.565025 54915 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.566536 54915 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.568354 54915 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.569832 54915 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.571398 54915 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.572876 54915 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.574387 54915 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.575824 54915 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? [PANICKED] Test Panicked In [AfterEach] at: /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/panic.go:262 @ 10/08/25 22:26:10.578 runtime error: invalid memory address or nil pointer dereference Full Stack Trace github.com/deckhouse/virtualization/tests/e2e.init.func21.3() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:76 +0xd4 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/08/25 22:26:10.578 (42ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/08/25 22:26:10.578 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:131 @ 10/08/25 22:26:10.578 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:177 @ 10/08/25 22:26:10.578 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:341 @ 10/08/25 22:26:10.578 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:367 @ 10/08/25 22:26:10.578 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-images-creation Unexpected error: <*exec.ExitError | 0x14000110000>: exit status 1 { ProcessState: { pid: 54916, status: 256, rusage: { Utime: { Sec: 0, Usec: 23248, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 8824, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45973504, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3610, Majflt: 4, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 46, Nvcsw: 22, Nivcsw: 512, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:48 @ 10/08/25 22:26:10.607 + > Enter [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:38 @ 10/08/25 22:26:10.578 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-images-creation Unexpected error: <*exec.ExitError | 0x14000110000>: exit status 1 { ProcessState: { pid: 54916, status: 256, rusage: { Utime: { Sec: 0, Usec: 23248, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 8824, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45973504, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3610, Majflt: 4, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 46, Nvcsw: 22, Nivcsw: 512, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:48 @ 10/08/25 22:26:10.607 < Exit [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:38 @ 10/08/25 22:26:10.607 (29ms) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/08/25 22:26:10.607 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=images-creation' error: exit status 1 stderr: E1008 22:26:10.633780 54917 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.635396 54917 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.636970 54917 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.638507 54917 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.640247 54917 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.641911 54917 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.643331 54917 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.644755 54917 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/08/25 22:26:10.647 (40ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:87 @ 10/08/25 22:26:10.647 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:96 @ 10/08/25 22:26:10.647 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:107 @ 10/08/25 22:26:10.647 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:116 @ 10/08/25 22:26:10.647 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:127 @ 10/08/25 22:26:10.647 + + + + > Enter [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/08/25 22:26:10.647 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:51 @ 10/08/25 22:26:10.647 < Exit [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/08/25 22:26:10.647 (0s) > Enter [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/08/25 22:26:10.647 < Exit [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/08/25 22:26:10.647 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:82 @ 10/08/25 22:26:10.647 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:94 @ 10/08/25 22:26:10.647 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:106 @ 10/08/25 22:26:10.647 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:116 @ 10/08/25 22:26:10.647 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:132 @ 10/08/25 22:26:10.647 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:151 @ 10/08/25 22:26:10.647 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:157 @ 10/08/25 22:26:10.647 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:189 @ 10/08/25 22:26:10.647 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:216 @ 10/08/25 22:26:10.647 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:252 @ 10/08/25 22:26:10.647 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-migration Unexpected error: <*exec.ExitError | 0x14000478260>: exit status 1 { ProcessState: { pid: 54918, status: 256, rusage: { Utime: { Sec: 0, Usec: 22006, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 9315, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44826624, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3541, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 44, Nvcsw: 25, Nivcsw: 618, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:42 @ 10/08/25 22:26:10.676 + > Enter [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/08/25 22:26:10.648 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-migration Unexpected error: <*exec.ExitError | 0x14000478260>: exit status 1 { ProcessState: { pid: 54918, status: 256, rusage: { Utime: { Sec: 0, Usec: 22006, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 9315, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44826624, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3541, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 44, Nvcsw: 25, Nivcsw: 618, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:42 @ 10/08/25 22:26:10.676 < Exit [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/08/25 22:26:10.676 (28ms) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/08/25 22:26:10.676 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-migration' error: exit status 1 stderr: E1008 22:26:10.703385 54919 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.705050 54919 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.706652 54919 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.708107 54919 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.709680 54919 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.711101 54919 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.712492 54919 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.713887 54919 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/08/25 22:26:10.716 (40ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/08/25 22:26:10.716 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:86 @ 10/08/25 22:26:10.716 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:100 @ 10/08/25 22:26:10.716 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:116 @ 10/08/25 22:26:10.716 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:131 @ 10/08/25 22:26:10.716 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-label-annotation Unexpected error: <*exec.ExitError | 0x140004784e0>: exit status 1 { ProcessState: { pid: 54920, status: 256, rusage: { Utime: { Sec: 0, Usec: 22187, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 9173, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45056000, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3559, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 66, Nvcsw: 33, Nivcsw: 590, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:48 @ 10/08/25 22:26:10.744 + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/08/25 22:26:10.716 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-label-annotation Unexpected error: <*exec.ExitError | 0x140004784e0>: exit status 1 { ProcessState: { pid: 54920, status: 256, rusage: { Utime: { Sec: 0, Usec: 22187, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 9173, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45056000, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3559, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 66, Nvcsw: 33, Nivcsw: 590, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:48 @ 10/08/25 22:26:10.744 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/08/25 22:26:10.744 (28ms) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/08/25 22:26:10.744 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-label-annotation' error: exit status 1 stderr: E1008 22:26:10.772462 54921 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.774058 54921 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.775727 54921 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.777230 54921 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.778731 54921 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.780262 54921 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.781718 54921 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.783239 54921 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/08/25 22:26:10.785 (41ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/08/25 22:26:10.785 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/08/25 22:26:10.785 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/08/25 22:26:10.786 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/08/25 22:26:10.786 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/08/25 22:26:10.786 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/08/25 22:26:10.786 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/08/25 22:26:10.786 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/08/25 22:26:10.786 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/08/25 22:26:10.786 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/08/25 22:26:10.786 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/08/25 22:26:10.786 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/08/25 22:26:10.786 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-sizing-policy Unexpected error: <*exec.ExitError | 0x14000479040>: exit status 1 { ProcessState: { pid: 54922, status: 256, rusage: { Utime: { Sec: 0, Usec: 22299, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 8753, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44695552, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3537, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 46, Nvcsw: 21, Nivcsw: 507, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:64 @ 10/08/25 22:26:10.814 + > Enter [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/08/25 22:26:10.786 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-sizing-policy Unexpected error: <*exec.ExitError | 0x14000479040>: exit status 1 { ProcessState: { pid: 54922, status: 256, rusage: { Utime: { Sec: 0, Usec: 22299, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 8753, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44695552, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3537, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 46, Nvcsw: 21, Nivcsw: 507, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:64 @ 10/08/25 22:26:10.814 < Exit [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/08/25 22:26:10.814 (28ms) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/08/25 22:26:10.814 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=sizing-policy' error: exit status 1 stderr: E1008 22:26:10.842134 54923 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.843686 54923 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.845314 54923 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.846851 54923 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.848396 54923 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.849850 54923 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.851377 54923 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.852833 54923 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/08/25 22:26:10.855 (40ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/08/25 22:26:10.855 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/08/25 22:26:10.855 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/08/25 22:26:10.855 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:126 @ 10/08/25 22:26:10.855 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:140 @ 10/08/25 22:26:10.855 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:152 @ 10/08/25 22:26:10.855 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:157 @ 10/08/25 22:26:10.855 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:163 @ 10/08/25 22:26:10.855 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:176 @ 10/08/25 22:26:10.855 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:181 @ 10/08/25 22:26:10.855 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:187 @ 10/08/25 22:26:10.855 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:202 @ 10/08/25 22:26:10.855 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:216 @ 10/08/25 22:26:10.855 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:240 @ 10/08/25 22:26:10.855 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-affinity-toleration Unexpected error: <*exec.ExitError | 0x14000502040>: exit status 1 { ProcessState: { pid: 54924, status: 256, rusage: { Utime: { Sec: 0, Usec: 21780, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 8332, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45121536, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3565, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 54, Nvcsw: 23, Nivcsw: 518, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:69 @ 10/08/25 22:26:10.884 + > Enter [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/08/25 22:26:10.855 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-affinity-toleration Unexpected error: <*exec.ExitError | 0x14000502040>: exit status 1 { ProcessState: { pid: 54924, status: 256, rusage: { Utime: { Sec: 0, Usec: 21780, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 8332, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45121536, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3565, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 54, Nvcsw: 23, Nivcsw: 518, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:69 @ 10/08/25 22:26:10.884 < Exit [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/08/25 22:26:10.884 (28ms) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/08/25 22:26:10.884 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=affinity-toleration' error: exit status 1 stderr: E1008 22:26:10.909598 54925 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.911149 54925 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.912651 54925 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.914142 54925 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.915597 54925 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.917045 54925 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.918541 54925 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.919984 54925 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/08/25 22:26:10.922 (38ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/08/25 22:26:10.922 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:120 @ 10/08/25 22:26:10.922 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:283 @ 10/08/25 22:26:10.922 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:372 @ 10/08/25 22:26:10.922 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:467 @ 10/08/25 22:26:10.922 + + + [FAILED] Unexpected error: <*exec.ExitError | 0x14000502440>: exit status 1 { ProcessState: { pid: 54926, status: 256, rusage: { Utime: { Sec: 0, Usec: 22319, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 9554, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45154304, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3566, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 45, Msgrcv: 45, Nsignals: 59, Nvcsw: 51, Nivcsw: 690, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:65 @ 10/08/25 22:26:10.959 + > Enter [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/08/25 22:26:10.922 [FAILED] Unexpected error: <*exec.ExitError | 0x14000502440>: exit status 1 { ProcessState: { pid: 54926, status: 256, rusage: { Utime: { Sec: 0, Usec: 22319, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 9554, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45154304, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3566, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 45, Msgrcv: 45, Nsignals: 59, Nvcsw: 51, Nivcsw: 690, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:65 @ 10/08/25 22:26:10.959 < Exit [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/08/25 22:26:10.959 (37ms) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/08/25 22:26:10.959 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=image-hotplug' error: exit status 1 stderr: E1008 22:26:10.987502 54927 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.989148 54927 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.990666 54927 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.992187 54927 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.993636 54927 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.995125 54927 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.996541 54927 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:10.997949 54927 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/08/25 22:26:11 (40ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:112 @ 10/08/25 22:26:11 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:147 @ 10/08/25 22:26:11 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:190 @ 10/08/25 22:26:11 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:196 @ 10/08/25 22:26:11 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:204 @ 10/08/25 22:26:11 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:224 @ 10/08/25 22:26:11 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:236 @ 10/08/25 22:26:11 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:258 @ 10/08/25 22:26:11 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:284 @ 10/08/25 22:26:11 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:294 @ 10/08/25 22:26:11 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:309 @ 10/08/25 22:26:11 + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:26:11 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11.001 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:26:11.001 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11.001 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11.001 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:26:11.001 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11.001 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11.001 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:26:11.001 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11.001 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11.001 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:26:11.001 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11.001 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11.001 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:26:11.001 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11.001 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11.001 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:26:11.001 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11.001 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11.001 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:26:11.001 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11.001 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11.002 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 22:26:11.002 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 22:26:11.002 (0s) + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-versions Unexpected error: <*exec.ExitError | 0x14000716e40>: exit status 1 { ProcessState: { pid: 54928, status: 256, rusage: { Utime: { Sec: 0, Usec: 22968, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 8432, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45285376, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3573, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 53, Nvcsw: 31, Nivcsw: 530, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:41 @ 10/08/25 22:26:11.031 + > Enter [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/08/25 22:26:11.002 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-versions Unexpected error: <*exec.ExitError | 0x14000716e40>: exit status 1 { ProcessState: { pid: 54928, status: 256, rusage: { Utime: { Sec: 0, Usec: 22968, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 8432, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45285376, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3573, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 53, Nvcsw: 31, Nivcsw: 530, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:41 @ 10/08/25 22:26:11.031 < Exit [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/08/25 22:26:11.031 (29ms) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/08/25 22:26:11.031 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-versions' error: exit status 1 stderr: E1008 22:26:11.058502 54929 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:11.060115 54929 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:11.061738 54929 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:11.063214 54929 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:11.064756 54929 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:11.066185 54929 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:11.067621 54929 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:11.069038 54929 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/08/25 22:26:11.071 (40ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/08/25 22:26:11.071 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/08/25 22:26:11.071 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/08/25 22:26:11.071 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/08/25 22:26:11.071 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/08/25 22:26:11.071 + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.071 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:26:11.071 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.071 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.071 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:26:11.071 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.071 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.071 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:26:11.071 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.071 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:26:11.072 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:26:11.072 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:26:11.072 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:26:11.072 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:26:11.072 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:26:11.072 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:26:11.072 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:26:11.072 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/08/25 22:26:11.072 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/08/25 22:26:11.072 (0s) + + + + > Enter [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/08/25 22:26:11.072 [SKIPPED] Module SDN is disabled. Skipping all tests for module SDN. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:50 @ 10/08/25 22:26:12.104 < Exit [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/08/25 22:26:12.104 (1.031s) > Enter [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/08/25 22:26:12.104 < Exit [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/08/25 22:26:12.104 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/08/25 22:26:12.104 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/08/25 22:26:12.104 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:111 @ 10/08/25 22:26:12.104 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:125 @ 10/08/25 22:26:12.104 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:141 @ 10/08/25 22:26:12.104 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:156 @ 10/08/25 22:26:12.104 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:169 @ 10/08/25 22:26:12.104 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-evacuation Unexpected error: <*exec.ExitError | 0x14000502040>: exit status 1 { ProcessState: { pid: 54930, status: 256, rusage: { Utime: { Sec: 0, Usec: 27944, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 12736, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44728320, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3541, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 52, Nvcsw: 25, Nivcsw: 609, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:48 @ 10/08/25 22:26:12.143 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/08/25 22:26:12.104 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-evacuation Unexpected error: <*exec.ExitError | 0x14000502040>: exit status 1 { ProcessState: { pid: 54930, status: 256, rusage: { Utime: { Sec: 0, Usec: 27944, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 12736, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44728320, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3541, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 52, Nvcsw: 25, Nivcsw: 609, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:48 @ 10/08/25 22:26:12.143 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/08/25 22:26:12.143 (38ms) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/08/25 22:26:12.143 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-evacuation' error: exit status 1 stderr: E1008 22:26:12.172014 54931 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.173719 54931 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.175239 54931 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.176656 54931 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.178182 54931 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.179613 54931 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.181100 54931 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.182507 54931 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/08/25 22:26:12.184 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/08/25 22:26:12.261 (77ms) [FAILED] cannot delete test case resources kustomizationDir: /tmp/testdata/vm-evacuation cmd: kubectl delete --kustomize /tmp/testdata/vm-evacuation --ignore-not-found stderr: E1008 22:26:12.245577 54932 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.247171 54932 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.248572 54932 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.250019 54932 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.251484 54932 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.252974 54932 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.254449 54932 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.255917 54932 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.257370 54932 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.258888 54932 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-evacuation": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused Unexpected error: <*exec.ExitError | 0x140006816e0>: exit status 1 { ProcessState: { pid: 54932, status: 256, rusage: { Utime: { Sec: 0, Usec: 82577, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 17914, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 77938688, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5572, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 63, Msgrcv: 63, Nsignals: 142, Nvcsw: 139, Nivcsw: 1250, }, }, Stderr: nil, } occurred In [AfterEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:617 @ 10/08/25 22:26:12.261 < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/08/25 22:26:12.261 (119ms) + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-restore-safe Unexpected error: <*exec.ExitError | 0x140001104a0>: exit status 1 { ProcessState: { pid: 54933, status: 256, rusage: { Utime: { Sec: 0, Usec: 23015, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 8228, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45236224, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3574, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 44, Nvcsw: 21, Nivcsw: 523, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:58 @ 10/08/25 22:26:12.289 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/08/25 22:26:12.261 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-restore-safe Unexpected error: <*exec.ExitError | 0x140001104a0>: exit status 1 { ProcessState: { pid: 54933, status: 256, rusage: { Utime: { Sec: 0, Usec: 23015, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 8228, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45236224, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3574, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 44, Nvcsw: 21, Nivcsw: 523, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:58 @ 10/08/25 22:26:12.289 < Exit [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/08/25 22:26:12.289 (28ms) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/08/25 22:26:12.289 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-restore-safe' error: exit status 1 stderr: E1008 22:26:12.316591 54934 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.318253 54934 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.319808 54934 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.321331 54934 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.322864 54934 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.324392 54934 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.325923 54934 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.327344 54934 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? [PANICKED] Test Panicked In [AfterEach] at: /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/panic.go:262 @ 10/08/25 22:26:12.329 runtime error: invalid memory address or nil pointer dereference Full Stack Trace github.com/deckhouse/virtualization/tests/e2e.init.func22.3() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:70 +0xd4 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/08/25 22:26:12.329 (40ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/08/25 22:26:12.329 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:115 @ 10/08/25 22:26:12.329 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:158 @ 10/08/25 22:26:12.329 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:348 @ 10/08/25 22:26:12.329 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:374 @ 10/08/25 22:26:12.329 + + + + > Enter [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/08/25 22:26:12.329 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:41 @ 10/08/25 22:26:12.329 < Exit [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/08/25 22:26:12.329 (0s) > Enter [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/08/25 22:26:12.329 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/08/25 22:26:12.329 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/08/25 22:26:12.405 (75ms) [FAILED] cannot delete test case resources kustomizationDir: /tmp/testdata/vm-migration-cancel cmd: kubectl delete --kustomize /tmp/testdata/vm-migration-cancel --ignore-not-found stderr: E1008 22:26:12.388994 54935 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.390604 54935 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.392119 54935 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.393592 54935 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.395054 54935 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.396493 54935 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.397907 54935 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.399375 54935 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.400866 54935 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.402361 54935 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/vm-migration-cancel": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused Unexpected error: <*exec.ExitError | 0x14000503240>: exit status 1 { ProcessState: { pid: 54935, status: 256, rusage: { Utime: { Sec: 0, Usec: 78147, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 17451, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 78315520, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5591, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 63, Msgrcv: 63, Nsignals: 141, Nvcsw: 157, Nivcsw: 1232, }, }, Stderr: nil, } occurred In [AfterEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:617 @ 10/08/25 22:26:12.405 < Exit [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/08/25 22:26:12.405 (75ms) + + + [FAILED] error: error validating "/tmp/testdata/importer-network-policy/project": error validating data: failed to download openapi: Get "http://localhost:8080/openapi/v2?timeout=32s": dial tcp [::1]:8080: connect: connection refused; if you choose to ignore these errors, turn validation off with --validate=false Expected <bool>: false to equal <bool>: true In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:69 @ 10/08/25 22:26:12.449 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/08/25 22:26:12.405 < Exit [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/08/25 22:26:12.405 (0s) > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/08/25 22:26:12.405 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/08/25 22:26:12.405 (0s) > Enter [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/08/25 22:26:12.405 [FAILED] error: error validating "/tmp/testdata/importer-network-policy/project": error validating data: failed to download openapi: Get "http://localhost:8080/openapi/v2?timeout=32s": dial tcp [::1]:8080: connect: connection refused; if you choose to ignore these errors, turn validation off with --validate=false Expected <bool>: false to equal <bool>: true In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:69 @ 10/08/25 22:26:12.449 < Exit [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/08/25 22:26:12.449 (43ms) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/08/25 22:26:12.449 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=importer-network-policy' error: exit status 1 stderr: E1008 22:26:12.476191 54938 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.477885 54938 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.479270 54938 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.480795 54938 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.482333 54938 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.483752 54938 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.485250 54938 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.486707 54938 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/08/25 22:26:12.489 (40ms) > Enter [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/08/25 22:26:12.489 STEP: Delete manifests - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:50 @ 10/08/25 22:26:12.489 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/08/25 22:26:12.489 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/08/25 22:26:12.529 (40ms) [FAILED] cannot delete test case resources kustomizationDir: /tmp/testdata/importer-network-policy cmd: kubectl delete --kustomize /tmp/testdata/importer-network-policy --ignore-not-found stderr: E1008 22:26:12.520974 54939 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.522496 54939 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.524020 54939 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.525456 54939 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.526992 54939 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" unable to recognize "/tmp/testdata/importer-network-policy": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/importer-network-policy": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/importer-network-policy": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/importer-network-policy": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused unable to recognize "/tmp/testdata/importer-network-policy": Get "http://localhost:8080/api?timeout=32s": dial tcp [::1]:8080: connect: connection refused Unexpected error: <*exec.ExitError | 0x140002920e0>: exit status 1 { ProcessState: { pid: 54939, status: 256, rusage: { Utime: { Sec: 0, Usec: 30486, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 10920, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 46841856, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3668, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 33, Msgrcv: 33, Nsignals: 66, Nvcsw: 53, Nivcsw: 803, }, }, Stderr: nil, } occurred In [AfterAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:617 @ 10/08/25 22:26:12.529 < Exit [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/08/25 22:26:12.529 (40ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/08/25 22:26:12.529 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/08/25 22:26:12.529 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/08/25 22:26:12.529 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/08/25 22:26:12.529 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/08/25 22:26:12.529 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-connectivity Unexpected error: <*exec.ExitError | 0x14000717ee0>: exit status 1 { ProcessState: { pid: 54940, status: 256, rusage: { Utime: { Sec: 0, Usec: 23400, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 9970, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45006848, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3556, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 63, Nvcsw: 47, Nivcsw: 532, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:63 @ 10/08/25 22:26:12.558 + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/08/25 22:26:12.529 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-connectivity Unexpected error: <*exec.ExitError | 0x14000717ee0>: exit status 1 { ProcessState: { pid: 54940, status: 256, rusage: { Utime: { Sec: 0, Usec: 23400, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 9970, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 45006848, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3556, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 63, Nvcsw: 47, Nivcsw: 532, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:63 @ 10/08/25 22:26:12.558 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/08/25 22:26:12.558 (29ms) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/08/25 22:26:12.558 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-connectivity' error: exit status 1 stderr: E1008 22:26:12.585630 54941 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.587240 54941 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.588710 54941 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.590121 54941 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.591563 54941 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.593364 54941 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.594894 54941 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.596423 54941 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/08/25 22:26:12.598 (40ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/08/25 22:26:12.598 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/08/25 22:26:12.598 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:118 @ 10/08/25 22:26:12.598 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:129 @ 10/08/25 22:26:12.598 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:142 @ 10/08/25 22:26:12.598 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:166 @ 10/08/25 22:26:12.598 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:174 @ 10/08/25 22:26:12.598 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:179 @ 10/08/25 22:26:12.598 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:187 @ 10/08/25 22:26:12.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:198 @ 10/08/25 22:26:12.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:209 @ 10/08/25 22:26:12.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:222 @ 10/08/25 22:26:12.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:228 @ 10/08/25 22:26:12.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:240 @ 10/08/25 22:26:12.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:250 @ 10/08/25 22:26:12.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:258 @ 10/08/25 22:26:12.599 + + + [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-disk-attachment Unexpected error: <*exec.ExitError | 0x14000d7c920>: exit status 1 { ProcessState: { pid: 54942, status: 256, rusage: { Utime: { Sec: 0, Usec: 21743, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 8048, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44908544, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3546, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 44, Nvcsw: 27, Nivcsw: 480, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:61 @ 10/08/25 22:26:12.626 + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/08/25 22:26:12.599 [FAILED] kubectl create namespace head-45bdb83d-end-to-end-vm-disk-attachment Unexpected error: <*exec.ExitError | 0x14000d7c920>: exit status 1 { ProcessState: { pid: 54942, status: 256, rusage: { Utime: { Sec: 0, Usec: 21743, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 8048, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 44908544, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3546, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 9, Msgrcv: 9, Nsignals: 44, Nvcsw: 27, Nivcsw: 480, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:61 @ 10/08/25 22:26:12.626 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/08/25 22:26:12.626 (28ms) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/08/25 22:26:12.626 Get resources error: cmd: kubectl get virtualization,intvirt,po,volumesnapshot -A -o yaml -l 'testcase=vm-disk-attachment' error: exit status 1 stderr: E1008 22:26:12.653686 54943 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.655405 54943 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.656928 54943 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.658410 54943 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.659881 54943 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.661336 54943 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.662755 54943 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:12.664223 54943 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/08/25 22:26:12.667 (40ms) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/08/25 22:26:12.667 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/08/25 22:26:12.667 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/08/25 22:26:12.667 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/08/25 22:26:12.667 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/08/25 22:26:12.667 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/08/25 22:26:12.667 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/08/25 22:26:12.667 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/08/25 22:26:12.667 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/08/25 22:26:12.667 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/08/25 22:26:12.667 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/08/25 22:26:12.667 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/08/25 22:26:12.667 + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400074f878>: the container "virtualization-controller" was not found: virtualization-controller-5f4b76784-4bs9l the container "virtualization-controller" was restarted: virtualization-controller-5f4b76784-f8s9l the container "virtualization-controller" was not found: virtualization-controller-68568fc8d-lph4f { errs: [ <*errors.joinError | 0x1400074f860>{ errs: [ <*errors.joinError | 0x1400074f848>{ errs: [ <*errors.errorString | 0x14000114c70>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-5f4b76784-4bs9l", }, ], }, <*errors.errorString | 0x14000114cd0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-5f4b76784-f8s9l", }, ], }, <*errors.errorString | 0x14000114d20>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-68568fc8d-lph4f", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/08/25 22:26:12.926 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/08/25 22:26:12.667 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/08/25 22:26:12.667 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/08/25 22:26:12.667 [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400074f878>: the container "virtualization-controller" was not found: virtualization-controller-5f4b76784-4bs9l the container "virtualization-controller" was restarted: virtualization-controller-5f4b76784-f8s9l the container "virtualization-controller" was not found: virtualization-controller-68568fc8d-lph4f { errs: [ <*errors.joinError | 0x1400074f860>{ errs: [ <*errors.joinError | 0x1400074f848>{ errs: [ <*errors.errorString | 0x14000114c70>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-5f4b76784-4bs9l", }, ], }, <*errors.errorString | 0x14000114cd0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-5f4b76784-f8s9l", }, ], }, <*errors.errorString | 0x14000114d20>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-68568fc8d-lph4f", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/08/25 22:26:12.926 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/08/25 22:26:12.926 (259ms) + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400057c018>: cmd: kubectl delete clustervirtualimages.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found stderr: E1008 22:26:13.073537 54944 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.075227 54944 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.076798 54944 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.078288 54944 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.079886 54944 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.081288 54944 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.082704 54944 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? cmd: kubectl delete virtualmachineclasses.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found stderr: E1008 22:26:13.113967 54945 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.115514 54945 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.117068 54945 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.118515 54945 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.119944 54945 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.121836 54945 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.123324 54945 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? cmd: kubectl delete virtualmachineipaddressleases.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found stderr: E1008 22:26:13.154184 54946 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.155687 54946 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.157264 54946 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.158751 54946 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.160227 54946 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.161803 54946 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.163423 54946 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? { errs: [ <*errors.joinError | 0x1400000e5b8>{ errs: [ <*errors.joinError | 0x1400000e408>{ errs: [ <*errors.errorString | 0x1400069c200>{ s: "cmd: kubectl delete clustervirtualimages.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found\nstderr: E1008 22:26:13.073537 54944 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.075227 54944 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.076798 54944 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.078288 54944 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.079886 54944 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.081288 54944 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.082704 54944 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nThe connection to the server localhost:8080 was refused - did you specify the right host or port?\n", }, ], }, <*errors.errorString | 0x1400069ccb0>{ s: "cmd: kubectl delete virtualmachineclasses.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found\nstderr: E1008 22:26:13.113967 54945 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.115514 54945 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.117068 54945 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.118515 54945 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.119944 54945 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.121836 54945 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.123324 54945 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nThe connection to the server localhost:8080 was refused - did you specify the right host or port?\n", }, ... Gomega truncated this representation as it exceeds 'format.MaxLength'. Consider having the object provide a custom 'GomegaStringer' representation or adjust the parameters in Gomega's 'format' package. Learn more here: https://onsi.github.io/gomega/#adjusting-output In [DeferCleanup (Suite)] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:177 @ 10/08/25 22:26:15.317 + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/08/25 22:26:12.926 [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400057c018>: cmd: kubectl delete clustervirtualimages.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found stderr: E1008 22:26:13.073537 54944 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.075227 54944 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.076798 54944 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.078288 54944 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.079886 54944 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.081288 54944 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.082704 54944 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? cmd: kubectl delete virtualmachineclasses.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found stderr: E1008 22:26:13.113967 54945 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.115514 54945 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.117068 54945 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.118515 54945 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.119944 54945 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.121836 54945 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.123324 54945 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? cmd: kubectl delete virtualmachineipaddressleases.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found stderr: E1008 22:26:13.154184 54946 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.155687 54946 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.157264 54946 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.158751 54946 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.160227 54946 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.161803 54946 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" E1008 22:26:13.163423 54946 memcache.go:265] "Unhandled Error" err="couldn't get current server API group list: Get \"http://localhost:8080/api?timeout=32s\": dial tcp [::1]:8080: connect: connection refused" The connection to the server localhost:8080 was refused - did you specify the right host or port? { errs: [ <*errors.joinError | 0x1400000e5b8>{ errs: [ <*errors.joinError | 0x1400000e408>{ errs: [ <*errors.errorString | 0x1400069c200>{ s: "cmd: kubectl delete clustervirtualimages.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found\nstderr: E1008 22:26:13.073537 54944 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.075227 54944 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.076798 54944 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.078288 54944 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.079886 54944 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.081288 54944 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.082704 54944 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nThe connection to the server localhost:8080 was refused - did you specify the right host or port?\n", }, ], }, <*errors.errorString | 0x1400069ccb0>{ s: "cmd: kubectl delete virtualmachineclasses.virtualization.deckhouse.io -l 'id=head-45bdb83d' --ignore-not-found\nstderr: E1008 22:26:13.113967 54945 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.115514 54945 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.117068 54945 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.118515 54945 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.119944 54945 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.121836 54945 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nE1008 22:26:13.123324 54945 memcache.go:265] \"Unhandled Error\" err=\"couldn't get current server API group list: Get \\\"http://localhost:8080/api?timeout=32s\\\": dial tcp [::1]:8080: connect: connection refused\"\nThe connection to the server localhost:8080 was refused - did you specify the right host or port?\n", }, ... Gomega truncated this representation as it exceeds 'format.MaxLength'. Consider having the object provide a custom 'GomegaStringer' representation or adjust the parameters in Gomega's 'format' package. Learn more here: https://onsi.github.io/gomega/#adjusting-output In [DeferCleanup (Suite)] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:177 @ 10/08/25 22:26:15.317 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/08/25 22:26:15.317 (2.39s) + + + \ No newline at end of file diff --git a/artifacts/m-sds-20251008-224318-11742/junit.xml b/artifacts/m-sds-20251008-224318-11742/junit.xml new file mode 100644 index 0000000000..50ad45cc4e --- /dev/null +++ b/artifacts/m-sds-20251008-224318-11742/junit.xml @@ -0,0 +1,678 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/08/25 23:05:46.294 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/08/25 23:05:50.374 (4.079s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/08/25 23:05:50.374 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/08/25 23:05:50.374 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.377 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 23:05:50.377 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.377 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.377 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 23:05:50.377 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.377 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.378 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 23:05:50.378 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.378 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.378 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 23:05:50.378 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.378 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.378 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 23:05:50.378 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.378 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.378 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 23:05:50.378 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.378 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.378 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 23:05:50.378 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.378 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.378 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 23:05:50.378 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.378 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.378 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/08/25 23:05:50.378 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/08/25 23:05:50.378 (0s) + + + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/08/25 23:05:50.378 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/08/25 23:05:52.423 (2.045s) > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/08/25 23:05:52.423 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/08/25 23:05:52.423 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/08/25 23:05:52.423 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/08/25 23:05:55.679 (3.256s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/08/25 23:05:55.679 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/08/25 23:05:55.679 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/08/25 23:05:55.68 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/08/25 23:05:55.68 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/08/25 23:05:55.68 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:75 @ 10/08/25 23:05:55.68 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/08/25 23:06:04.227 (8.547s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/08/25 23:06:04.227 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/08/25 23:06:04.227 (0s) + + + [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/08/25 23:10:39.16 This is the Progress Report generated when the suite timeout occurred: VirtualMachineLabelAndAnnotation When virtual disks are applied checks VDs phases (Spec Runtime: 4m34.95s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 In [It] (Node Runtime: 4m34.95s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 At [By Step] VDs should be in Ready phases (Step Runtime: 4m34.95s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 Spec Goroutine goroutine 184 [sync.WaitGroup.Wait, 4 minutes] sync.runtime_SemacquireWaitGroup(0x102f58b30?, 0xe0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x140008e4370) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000790c20, 0x1, 0x1049c8e70?}, {0x1046e28ae, 0x28}, {{0x0, 0x0, 0x0}, 0x140005420c0, {0x140009ab740, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1046e28ae, 0x28}, {{0x0, 0x0, 0x0}, 0x140005420c0, {0x140009ab740, 0x2c}, {0x14000792240, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1046e28ae, 0x28}, {0x1046a1865, 0x5}, {{0x0, 0x0, 0x0}, 0x140005420c0, {0x140009ab740, 0x2c}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func18.6.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phases", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 47 [syscall, 4 minutes] syscall.syscall6(0x1008e43b0?, 0x1072bd110?, 0x1072b0a78?, 0x90?, 0x140000a5808?, 0x1400022a510?, 0x14000622a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000622a98?, 0x102fda6fc?, 0x90?, 0x104fc8660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400028a540?, 0x14000622ad4, 0x140008e43b0?, 0x1400028a3f0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x1400044e580) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000915408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000199800) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000199800) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002b79a0?, 0x1050601e8?, 0x1400028a2a0?}}, {0x105080808?, 0x1400028a2a0?}, {0x14000938000?, 0xc231cfea54b27e10?}, {0x105059120, 0x1400044e500}, {0x105059120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002b79a0?, 0x106567668?, 0x10653dee0?}}, {0x105080808, 0x1400028a2a0}, {0x14000938000, 0xcd}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x105084e48?, 0x14000427d58?}, {0x1046a4924?, 0x1400084e6d8?}}, {0x1046e28ae, 0x28}, {0x14000a94900, 0x26}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 184 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 130 [sync.Cond.Wait, 4 minutes] sync.runtime_notifyListWait(0x14000778048, 0x2e) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x14000778038) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x14000778030, {0x1400077d000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x14000542d80?}, {0x1400077d000?, 0x104b35640?, 0x14000afe000?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x14000844000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x14000844000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000992008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000992008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000992008, {0x14000ba16a3, 0xf295d, 0x10309dc74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400028c008, {0x14000ba16a3, 0xf295d, 0xf295d}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000ba16a3?, 0x14000af9228?, 0x14000ba154b?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x1400085bda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140009801e0, {0x1129feb18, 0x140009802d0}, {0x105059920, 0x14000119380}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 86 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/08/25 23:06:04.227 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/08/25 23:06:04.227 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/08/25 23:06:04.227 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 @ 10/08/25 23:06:04.227 [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/08/25 23:10:39.16 This is the Progress Report generated when the suite timeout occurred: VirtualMachineLabelAndAnnotation When virtual disks are applied checks VDs phases (Spec Runtime: 4m34.95s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 In [It] (Node Runtime: 4m34.95s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 At [By Step] VDs should be in Ready phases (Step Runtime: 4m34.95s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 Spec Goroutine goroutine 184 [sync.WaitGroup.Wait, 4 minutes] sync.runtime_SemacquireWaitGroup(0x102f58b30?, 0xe0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x140008e4370) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000790c20, 0x1, 0x1049c8e70?}, {0x1046e28ae, 0x28}, {{0x0, 0x0, 0x0}, 0x140005420c0, {0x140009ab740, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1046e28ae, 0x28}, {{0x0, 0x0, 0x0}, 0x140005420c0, {0x140009ab740, 0x2c}, {0x14000792240, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1046e28ae, 0x28}, {0x1046a1865, 0x5}, {{0x0, 0x0, 0x0}, 0x140005420c0, {0x140009ab740, 0x2c}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func18.6.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phases", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 47 [syscall, 4 minutes] syscall.syscall6(0x1008e43b0?, 0x1072bd110?, 0x1072b0a78?, 0x90?, 0x140000a5808?, 0x1400022a510?, 0x14000622a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000622a98?, 0x102fda6fc?, 0x90?, 0x104fc8660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400028a540?, 0x14000622ad4, 0x140008e43b0?, 0x1400028a3f0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x1400044e580) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000915408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000199800) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000199800) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002b79a0?, 0x1050601e8?, 0x1400028a2a0?}}, {0x105080808?, 0x1400028a2a0?}, {0x14000938000?, 0xc231cfea54b27e10?}, {0x105059120, 0x1400044e500}, {0x105059120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002b79a0?, 0x106567668?, 0x10653dee0?}}, {0x105080808, 0x1400028a2a0}, {0x14000938000, 0xcd}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x105084e48?, 0x14000427d58?}, {0x1046a4924?, 0x1400084e6d8?}}, {0x1046e28ae, 0x28}, {0x14000a94900, 0x26}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 184 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 130 [sync.Cond.Wait, 4 minutes] sync.runtime_notifyListWait(0x14000778048, 0x2e) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x14000778038) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x14000778030, {0x1400077d000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x14000542d80?}, {0x1400077d000?, 0x104b35640?, 0x14000afe000?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x14000844000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x14000844000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000992008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000992008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000992008, {0x14000ba16a3, 0xf295d, 0x10309dc74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400028c008, {0x14000ba16a3, 0xf295d, 0xf295d}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000ba16a3?, 0x14000af9228?, 0x14000ba154b?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x1400085bda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140009801e0, {0x1129feb18, 0x140009802d0}, {0x105059920, 0x14000119380}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 86 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/08/25 23:10:39.165 (4m34.955s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/08/25 23:10:39.165 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/08/25 23:10:44.274 (5.11s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x140009ac300>: the container "virtualization-controller" was restarted: virtualization-controller-5cc678446-bs2nn the container "virtualization-controller" was not found: virtualization-controller-77d6d5cf5d-2djfl { errs: [ <*errors.joinError | 0x140009ac2e8>{ errs: [ <*errors.errorString | 0x14000886440>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-5cc678446-bs2nn", }, ], }, <*errors.errorString | 0x14000886470>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-77d6d5cf5d-2djfl", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/08/25 23:10:44.554 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/08/25 23:10:44.277 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/08/25 23:10:44.277 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/08/25 23:10:44.277 [FAILED] Expected success, but got an error: <*errors.joinError | 0x140009ac300>: the container "virtualization-controller" was restarted: virtualization-controller-5cc678446-bs2nn the container "virtualization-controller" was not found: virtualization-controller-77d6d5cf5d-2djfl { errs: [ <*errors.joinError | 0x140009ac2e8>{ errs: [ <*errors.errorString | 0x14000886440>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-5cc678446-bs2nn", }, ], }, <*errors.errorString | 0x14000886470>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-77d6d5cf5d-2djfl", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/08/25 23:10:44.554 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/08/25 23:10:44.554 (276ms) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/08/25 23:10:44.554 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/08/25 23:10:48.24 (3.686s) + + + \ No newline at end of file diff --git a/artifacts/matrix-sds-20251014-154647/junit.xml b/artifacts/matrix-sds-20251014-154647/junit.xml new file mode 100644 index 0000000000..d8a633a96f --- /dev/null +++ b/artifacts/matrix-sds-20251014-154647/junit.xml @@ -0,0 +1,2 @@ + + \ No newline at end of file diff --git a/artifacts/quick-cephrbd-20251009-144829-18312/junit.xml b/artifacts/quick-cephrbd-20251009-144829-18312/junit.xml new file mode 100644 index 0000000000..e44498428c --- /dev/null +++ b/artifacts/quick-cephrbd-20251009-144829-18312/junit.xml @@ -0,0 +1,669 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 16:10:24.404 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 16:10:29.156 (4.745s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 16:10:29.156 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 16:10:29.156 (0s) + + + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/09/25 16:10:29.159 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/09/25 16:10:30.681 (1.52s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/09/25 16:10:30.681 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/09/25 16:10:35.279 (4.593s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/09/25 16:10:35.279 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/09/25 16:10:35.279 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/09/25 16:10:35.279 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:97 @ 10/09/25 16:10:35.279 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/09/25 16:10:53.973 (18.684s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/09/25 16:10:53.973 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/09/25 16:10:53.973 (0s) + + + [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/09/25 16:20:24.031 This is the Progress Report generated when the suite timeout occurred: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 9m30.047s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [It] (Node Runtime: 9m30.047s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 At [By Step] VDs should be in Ready phase (Step Runtime: 9m30.047s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 Spec Goroutine goroutine 151 [sync.WaitGroup.Wait, 9 minutes] sync.runtime_SemacquireWaitGroup(0x100e1cb30?, 0xb0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x1400059e8d0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000140020, 0x2, 0x10288ce70?}, {0x1025a68ae, 0x28}, {{0x0, 0x0, 0x0}, 0x140005731d0, {0x1400098a750, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1025a68ae, 0x28}, {{0x0, 0x0, 0x0}, 0x140005731d0, {0x1400098a750, 0x25}, {0x1400078e4e0, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1025a68ae, 0x28}, {0x102565865, 0x5}, {{0x0, 0x0, 0x0}, 0x140005731d0, {0x1400098a750, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x140005a64e0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 10 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 120 [sync.Cond.Wait, 9 minutes] sync.runtime_notifyListWait(0x14000c8a048, 0x1f) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x14000c8a038) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x14000c8a030, {0x14000af3000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x140009ed410?}, {0x14000af3000?, 0x1029f9640?, 0x140003fe160?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x14000880060) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x14000880060) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x140007f1308) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x140007f1308) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x140007f1308, {0x14000ca5832, 0xf87ce, 0x100f61c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400034a848, {0x14000ca5832, 0xf87ce, 0xf87ce}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000ca5832?, 0x1400004e928?, 0x14000ca55bc?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x1400093bda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x1400053d410, {0x130999598, 0x140005a62a0}, {0x102f1d920, 0x14000077920}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 67 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 70 [syscall, 9 minutes] syscall.syscall6(0x10059ed80?, 0x13071bf80?, 0x105174f30?, 0x90?, 0x14000980008?, 0x14000868120?, 0x140005c3a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140005c3a98?, 0x100e9e6fc?, 0x90?, 0x102e8c660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002a8230?, 0x140005c3ad4, 0x1400059ed80?, 0x140002a81c0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140006fc180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140008ce408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000197500) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000197500) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140001799a0?, 0x102f241e8?, 0x140002a80e0?}}, {0x102f44808?, 0x140002a80e0?}, {0x140008c6000?, 0xc2320bf6c7b5ffc0?}, {0x102f1d120, 0x140006fc100}, {0x102f1d120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140001799a0?, 0x10442b668?, 0x104401ee0?}}, {0x102f44808, 0x140002a80e0}, {0x140008c6000, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102f48e48?, 0x14000494720?}, {0x102568924?, 0x140005c3e98?}}, {0x1025a68ae, 0x28}, {0x14000454220, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 151 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 111 [sync.Cond.Wait] sync.runtime_notifyListWait(0x140009e0ac8, 0x45) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140009e0ab8) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140009e0ab0, {0x14000af2000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x140009ed410?}, {0x14000af2000?, 0x1029f9640?, 0x140005fa0b0?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x1400080b5c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x1400080b5c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x140009cb908) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x140009cb908) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x140009cb908, {0x14000b0d522, 0xf2ade, 0x100f61c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400028e848, {0x14000b0d522, 0xf2ade, 0xf2ade}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000b0d522?, 0x140009f00b8?, 0x14000b0d355?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x1400093dda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x1400053d410, {0x130999598, 0x140009ed200}, {0x102f1d920, 0x14000077920}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 67 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 69 [syscall, 9 minutes] syscall.syscall6(0x100011010?, 0x13071c3e0?, 0x105174108?, 0x90?, 0x14000601008?, 0x14000c9c5a0?, 0x140008fba68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140008fba98?, 0x100e9e6fc?, 0x90?, 0x102e8c660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029afc0?, 0x140008fbad4, 0x14000011010?, 0x1400029af50?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140002a6f00) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000a55008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000c8a600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000c8a600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140001799a0?, 0x102f241e8?, 0x1400029ae70?}}, {0x102f44808?, 0x1400029ae70?}, {0x140004fc600?, 0xc2320bf6c7b5ffc0?}, {0x102f1d120, 0x140002a6c40}, {0x102f1d120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140001799a0?, 0x10442b668?, 0x104401ee0?}}, {0x102f44808, 0x1400029ae70}, {0x140004fc600, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102f48e48?, 0x14000494720?}, {0x102568924?, 0x102588660?}}, {0x1025a68ae, 0x28}, {0x14000454200, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 151 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) + > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/09/25 16:10:53.974 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 @ 10/09/25 16:10:53.974 [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/09/25 16:20:24.031 This is the Progress Report generated when the suite timeout occurred: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 9m30.047s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [It] (Node Runtime: 9m30.047s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 At [By Step] VDs should be in Ready phase (Step Runtime: 9m30.047s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 Spec Goroutine goroutine 151 [sync.WaitGroup.Wait, 9 minutes] sync.runtime_SemacquireWaitGroup(0x100e1cb30?, 0xb0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x1400059e8d0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000140020, 0x2, 0x10288ce70?}, {0x1025a68ae, 0x28}, {{0x0, 0x0, 0x0}, 0x140005731d0, {0x1400098a750, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1025a68ae, 0x28}, {{0x0, 0x0, 0x0}, 0x140005731d0, {0x1400098a750, 0x25}, {0x1400078e4e0, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1025a68ae, 0x28}, {0x102565865, 0x5}, {{0x0, 0x0, 0x0}, 0x140005731d0, {0x1400098a750, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x140005a64e0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 10 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 120 [sync.Cond.Wait, 9 minutes] sync.runtime_notifyListWait(0x14000c8a048, 0x1f) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x14000c8a038) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x14000c8a030, {0x14000af3000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x140009ed410?}, {0x14000af3000?, 0x1029f9640?, 0x140003fe160?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x14000880060) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x14000880060) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x140007f1308) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x140007f1308) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x140007f1308, {0x14000ca5832, 0xf87ce, 0x100f61c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400034a848, {0x14000ca5832, 0xf87ce, 0xf87ce}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000ca5832?, 0x1400004e928?, 0x14000ca55bc?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x1400093bda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x1400053d410, {0x130999598, 0x140005a62a0}, {0x102f1d920, 0x14000077920}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 67 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 70 [syscall, 9 minutes] syscall.syscall6(0x10059ed80?, 0x13071bf80?, 0x105174f30?, 0x90?, 0x14000980008?, 0x14000868120?, 0x140005c3a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140005c3a98?, 0x100e9e6fc?, 0x90?, 0x102e8c660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002a8230?, 0x140005c3ad4, 0x1400059ed80?, 0x140002a81c0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140006fc180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140008ce408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000197500) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000197500) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140001799a0?, 0x102f241e8?, 0x140002a80e0?}}, {0x102f44808?, 0x140002a80e0?}, {0x140008c6000?, 0xc2320bf6c7b5ffc0?}, {0x102f1d120, 0x140006fc100}, {0x102f1d120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140001799a0?, 0x10442b668?, 0x104401ee0?}}, {0x102f44808, 0x140002a80e0}, {0x140008c6000, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102f48e48?, 0x14000494720?}, {0x102568924?, 0x140005c3e98?}}, {0x1025a68ae, 0x28}, {0x14000454220, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 151 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 111 [sync.Cond.Wait] sync.runtime_notifyListWait(0x140009e0ac8, 0x45) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140009e0ab8) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140009e0ab0, {0x14000af2000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x140009ed410?}, {0x14000af2000?, 0x1029f9640?, 0x140005fa0b0?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x1400080b5c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x1400080b5c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x140009cb908) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x140009cb908) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x140009cb908, {0x14000b0d522, 0xf2ade, 0x100f61c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400028e848, {0x14000b0d522, 0xf2ade, 0xf2ade}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000b0d522?, 0x140009f00b8?, 0x14000b0d355?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x1400093dda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x1400053d410, {0x130999598, 0x140009ed200}, {0x102f1d920, 0x14000077920}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 67 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 69 [syscall, 9 minutes] syscall.syscall6(0x100011010?, 0x13071c3e0?, 0x105174108?, 0x90?, 0x14000601008?, 0x14000c9c5a0?, 0x140008fba68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140008fba98?, 0x100e9e6fc?, 0x90?, 0x102e8c660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029afc0?, 0x140008fbad4, 0x14000011010?, 0x1400029af50?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140002a6f00) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000a55008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000c8a600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000c8a600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140001799a0?, 0x102f241e8?, 0x1400029ae70?}}, {0x102f44808?, 0x1400029ae70?}, {0x140004fc600?, 0xc2320bf6c7b5ffc0?}, {0x102f1d120, 0x140002a6c40}, {0x102f1d120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140001799a0?, 0x10442b668?, 0x104401ee0?}}, {0x102f44808, 0x1400029ae70}, {0x140004fc600, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102f48e48?, 0x14000494720?}, {0x102568924?, 0x102588660?}}, {0x1025a68ae, 0x28}, {0x14000454200, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 151 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/09/25 16:20:24.036 (9m30.052s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/09/25 16:20:24.036 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/09/25 16:20:29.029 (4.993s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400080d458>: the container "virtualization-controller" was not found: virtualization-controller-cb4bf777b-klfx7 { errs: [ <*errors.errorString | 0x14000664430>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-cb4bf777b-klfx7", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/09/25 16:20:29.391 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 16:20:29.033 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 16:20:29.033 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 16:20:29.033 [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400080d458>: the container "virtualization-controller" was not found: virtualization-controller-cb4bf777b-klfx7 { errs: [ <*errors.errorString | 0x14000664430>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-cb4bf777b-klfx7", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/09/25 16:20:29.391 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 16:20:29.391 (358ms) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/09/25 16:20:29.392 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/09/25 16:20:33.139 (3.748s) + + + \ No newline at end of file diff --git a/artifacts/quick-sds-20251009-115641-10566/junit.xml b/artifacts/quick-sds-20251009-115641-10566/junit.xml new file mode 100644 index 0000000000..89a6079d2e --- /dev/null +++ b/artifacts/quick-sds-20251009-115641-10566/junit.xml @@ -0,0 +1,705 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 12:19:45.251 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 12:19:50.185 (4.934s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 12:19:50.185 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 12:19:50.185 (0s) + + + + > Enter [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/09/25 12:19:50.188 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:51 @ 10/09/25 12:19:50.189 < Exit [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/09/25 12:19:50.189 (1ms) > Enter [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/09/25 12:19:50.189 < Exit [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/09/25 12:19:50.189 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:82 @ 10/09/25 12:19:50.189 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:94 @ 10/09/25 12:19:50.189 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:106 @ 10/09/25 12:19:50.189 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:116 @ 10/09/25 12:19:50.189 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:132 @ 10/09/25 12:19:50.189 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:151 @ 10/09/25 12:19:50.19 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:157 @ 10/09/25 12:19:50.19 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:189 @ 10/09/25 12:19:50.19 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:216 @ 10/09/25 12:19:50.19 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:252 @ 10/09/25 12:19:50.19 + + + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:51 @ 10/09/25 12:19:50.19 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:51 @ 10/09/25 12:19:52.19 (2s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:61 @ 10/09/25 12:19:52.19 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:61 @ 10/09/25 12:20:22.334 (30.143s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 12:20:22.334 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 12:20:22.334 (0s) + + + > Enter [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:82 @ 10/09/25 12:20:22.335 < Exit [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:82 @ 10/09/25 12:20:23.838 (1.503s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 12:20:23.838 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 12:20:23.838 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:97 @ 10/09/25 12:20:23.839 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:98 @ 10/09/25 12:20:23.839 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:97 @ 10/09/25 12:20:25.314 (1.476s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 12:20:25.314 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 12:20:25.314 (0s) + + + > Enter [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:108 @ 10/09/25 12:20:25.315 STEP: CVIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:109 @ 10/09/25 12:20:25.315 < Exit [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:108 @ 10/09/25 12:20:26.369 (1.054s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 12:20:26.369 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 12:20:26.369 (0s) + + + > Enter [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:119 @ 10/09/25 12:20:26.369 STEP: VMClasses should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:120 @ 10/09/25 12:20:26.369 < Exit [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:119 @ 10/09/25 12:20:27.789 (1.42s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 12:20:27.789 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 12:20:27.789 (0s) + + + [FAILED] Timed out after 345.812s. Expected success, but got an error: <*errors.errorString | 0x140004da0b0>: cannot patch VMIP "head-45bdb83d-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-45bdb83d-vm-custom-ip { s: "cannot patch VMIP \"head-45bdb83d-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-45bdb83d-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/09/25 12:26:13.603 + > Enter [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 @ 10/09/25 12:20:27.79 [FAILED] Timed out after 345.812s. Expected success, but got an error: <*errors.errorString | 0x140004da0b0>: cannot patch VMIP "head-45bdb83d-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-45bdb83d-vm-custom-ip { s: "cannot patch VMIP \"head-45bdb83d-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-45bdb83d-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/09/25 12:26:13.603 < Exit [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 @ 10/09/25 12:26:13.604 (5m45.814s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 12:26:13.604 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 12:26:18.42 (4.816s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:137 @ 10/09/25 12:26:18.42 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:148 @ 10/09/25 12:26:18.42 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:158 @ 10/09/25 12:26:18.42 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:169 @ 10/09/25 12:26:18.42 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:180 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:192 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:216 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:240 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:263 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:275 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:293 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:310 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:334 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:347 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:351 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:368 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:381 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:398 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:437 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:472 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:490 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:505 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:521 @ 10/09/25 12:26:18.421 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:538 @ 10/09/25 12:26:18.422 + + + [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/09/25 12:29:37.515 This is the Progress Report generated when the suite timeout occurred: VirtualMachineEvacuation Evacuation (Spec Runtime: 3m19.058s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 In [It] (Node Runtime: 3m11.873s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 At [By Step] Virtual machine agents should be ready (Step Runtime: 3m11.873s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 Spec Goroutine goroutine 249 [sync.WaitGroup.Wait, 3 minutes] sync.runtime_SemacquireWaitGroup(0x102534b30?, 0xd0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x1400130c160) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140000de320, 0x2, 0x103fa4e70?}, {0x103cc30d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400061f6b0, {0x1400130af30, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x103cc30d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400061f6b0, {0x1400130af30, 0x26}, {0x1400130b3b0, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x103cc30d7, 0x2a}, {0x103c8037b, 0x7}, {{0x0, 0x0, 0x0}, 0x1400061f6b0, {0x1400130af30, 0x26}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x1400061f6b0, {0x1400130af30, 0x26}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func17.5() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 | It("Evacuation", func() { | By("Virtual machine agents should be ready") > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 26 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 260 [syscall, 3 minutes] syscall.syscall6(0x100a8c350?, 0x10b138b70?, 0x10b12ca78?, 0x90?, 0x14000688008?, 0x14000065560?, 0x1400091ca68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400091ca98?, 0x1025b66fc?, 0x90?, 0x1045a4660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400026b500?, 0x1400091cad4, 0x14000a8c350?, 0x1400026b490?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000b11880) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x1400078b408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000986600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000986600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a2000?, 0x10463c1e8?, 0x1400026b3b0?}}, {0x10465c808?, 0x1400026b3b0?}, {0x140004ae0d0?, 0xc231fecfb6b50910?}, {0x104635120, 0x14000b11800}, {0x104635120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a2000?, 0x105b43668?, 0x105b19ee0?}}, {0x10465c808, 0x1400026b3b0}, {0x140004ae0d0, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x104660e48?, 0x1400000e018?}, {0x103c80924?, 0x0?}}, {0x103cc30d7, 0x2a}, {0x14000632380, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 249 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 261 [syscall, 3 minutes] syscall.syscall6(0x10130c1a0?, 0x152057430?, 0x10b12c5c0?, 0x90?, 0x140000a7008?, 0x1400097e360?, 0x1400059ba68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400059ba98?, 0x1025b66fc?, 0x90?, 0x1045a4660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400019c7e0?, 0x1400059bad4, 0x1400130c1a0?, 0x1400019c770?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000117800) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000051c08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000171680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000171680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a2000?, 0x10463c1e8?, 0x1400019c5b0?}}, {0x10465c808?, 0x1400019c5b0?}, {0x140006fa000?, 0xc231fecfb6b46118?}, {0x104635120, 0x140001176c0}, {0x104635120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a2000?, 0x105b43668?, 0x105b19ee0?}}, {0x10465c808, 0x1400019c5b0}, {0x140006fa000, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x104660e48?, 0x1400000e018?}, {0x103c80924?, 0x1400059be98?}}, {0x103cc30d7, 0x2a}, {0x140006323a0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 249 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/09/25 12:26:18.422 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/09/25 12:26:20.396 (1.974s) > Enter [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/09/25 12:26:20.396 < Exit [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/09/25 12:26:25.607 (5.21s) > Enter [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/09/25 12:26:25.607 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 @ 10/09/25 12:26:25.607 [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/09/25 12:29:37.515 This is the Progress Report generated when the suite timeout occurred: VirtualMachineEvacuation Evacuation (Spec Runtime: 3m19.058s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 In [It] (Node Runtime: 3m11.873s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 At [By Step] Virtual machine agents should be ready (Step Runtime: 3m11.873s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 Spec Goroutine goroutine 249 [sync.WaitGroup.Wait, 3 minutes] sync.runtime_SemacquireWaitGroup(0x102534b30?, 0xd0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x1400130c160) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140000de320, 0x2, 0x103fa4e70?}, {0x103cc30d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400061f6b0, {0x1400130af30, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x103cc30d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400061f6b0, {0x1400130af30, 0x26}, {0x1400130b3b0, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x103cc30d7, 0x2a}, {0x103c8037b, 0x7}, {{0x0, 0x0, 0x0}, 0x1400061f6b0, {0x1400130af30, 0x26}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x1400061f6b0, {0x1400130af30, 0x26}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func17.5() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 | It("Evacuation", func() { | By("Virtual machine agents should be ready") > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 26 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 260 [syscall, 3 minutes] syscall.syscall6(0x100a8c350?, 0x10b138b70?, 0x10b12ca78?, 0x90?, 0x14000688008?, 0x14000065560?, 0x1400091ca68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400091ca98?, 0x1025b66fc?, 0x90?, 0x1045a4660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400026b500?, 0x1400091cad4, 0x14000a8c350?, 0x1400026b490?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000b11880) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x1400078b408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000986600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000986600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a2000?, 0x10463c1e8?, 0x1400026b3b0?}}, {0x10465c808?, 0x1400026b3b0?}, {0x140004ae0d0?, 0xc231fecfb6b50910?}, {0x104635120, 0x14000b11800}, {0x104635120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a2000?, 0x105b43668?, 0x105b19ee0?}}, {0x10465c808, 0x1400026b3b0}, {0x140004ae0d0, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x104660e48?, 0x1400000e018?}, {0x103c80924?, 0x0?}}, {0x103cc30d7, 0x2a}, {0x14000632380, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 249 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 261 [syscall, 3 minutes] syscall.syscall6(0x10130c1a0?, 0x152057430?, 0x10b12c5c0?, 0x90?, 0x140000a7008?, 0x1400097e360?, 0x1400059ba68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400059ba98?, 0x1025b66fc?, 0x90?, 0x1045a4660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400019c7e0?, 0x1400059bad4, 0x1400130c1a0?, 0x1400019c770?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000117800) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000051c08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000171680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000171680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a2000?, 0x10463c1e8?, 0x1400019c5b0?}}, {0x10465c808?, 0x1400019c5b0?}, {0x140006fa000?, 0xc231fecfb6b46118?}, {0x104635120, 0x140001176c0}, {0x104635120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a2000?, 0x105b43668?, 0x105b19ee0?}}, {0x10465c808, 0x1400019c5b0}, {0x140006fa000, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x104660e48?, 0x1400000e018?}, {0x103c80924?, 0x1400059be98?}}, {0x103cc30d7, 0x2a}, {0x140006323a0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 249 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/09/25 12:29:37.521 (3m11.88s) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/09/25 12:29:37.521 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/09/25 12:29:42.876 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/09/25 12:29:56.326 (13.45s) < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/09/25 12:29:56.326 (18.805s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x14000405968>: the container "virtualization-controller" was restarted: virtualization-controller-6594549467-h48mn the container "virtualization-controller" was restarted: virtualization-controller-6594549467-tm9qp the container "virtualization-controller" was not found: virtualization-controller-978f84689-5clqz { errs: [ <*errors.joinError | 0x14000405938>{ errs: [ <*errors.joinError | 0x14000405908>{ errs: [ <*errors.errorString | 0x1400064e5a0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-6594549467-h48mn", }, ], }, <*errors.errorString | 0x1400064e5d0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-6594549467-tm9qp", }, ], }, <*errors.errorString | 0x1400064e5f0>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-978f84689-5clqz", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/09/25 12:29:56.966 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 12:29:56.33 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 12:29:56.33 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 12:29:56.33 [FAILED] Expected success, but got an error: <*errors.joinError | 0x14000405968>: the container "virtualization-controller" was restarted: virtualization-controller-6594549467-h48mn the container "virtualization-controller" was restarted: virtualization-controller-6594549467-tm9qp the container "virtualization-controller" was not found: virtualization-controller-978f84689-5clqz { errs: [ <*errors.joinError | 0x14000405938>{ errs: [ <*errors.joinError | 0x14000405908>{ errs: [ <*errors.errorString | 0x1400064e5a0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-6594549467-h48mn", }, ], }, <*errors.errorString | 0x1400064e5d0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-6594549467-tm9qp", }, ], }, <*errors.errorString | 0x1400064e5f0>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-978f84689-5clqz", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/09/25 12:29:56.966 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 12:29:56.966 (636ms) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/09/25 12:29:56.966 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/09/25 12:30:09.456 (12.49s) + + + \ No newline at end of file diff --git a/artifacts/quick-sds-20251009-144829-12407/junit.xml b/artifacts/quick-sds-20251009-144829-12407/junit.xml new file mode 100644 index 0000000000..4195d3e8fc --- /dev/null +++ b/artifacts/quick-sds-20251009-144829-12407/junit.xml @@ -0,0 +1,669 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 16:09:50.255 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 16:09:55.057 (4.802s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 16:09:55.057 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 16:09:55.057 (0s) + + + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/09/25 16:09:55.063 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/09/25 16:09:57.031 (1.968s) > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/09/25 16:09:57.031 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/09/25 16:09:57.031 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/09/25 16:09:57.031 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/09/25 16:10:00.342 (3.311s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/09/25 16:10:00.342 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/09/25 16:10:00.342 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/09/25 16:10:00.343 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/09/25 16:10:00.343 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/09/25 16:10:00.343 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:75 @ 10/09/25 16:10:00.343 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/09/25 16:10:08.283 (7.941s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/09/25 16:10:08.283 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/09/25 16:10:08.284 (0s) + + + [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/09/25 16:19:43.188 This is the Progress Report generated when the suite timeout occurred: VirtualMachineLabelAndAnnotation When virtual disks are applied checks VDs phases (Spec Runtime: 9m34.864s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 In [It] (Node Runtime: 9m34.863s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 At [By Step] VDs should be in Ready phases (Step Runtime: 9m34.863s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 Spec Goroutine goroutine 148 [sync.WaitGroup.Wait, 9 minutes] sync.runtime_SemacquireWaitGroup(0x1046fcb30?, 0x60?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x1400026ebe0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140005a0ff0, 0x1, 0x10616ce70?}, {0x105e868ae, 0x28}, {{0x0, 0x0, 0x0}, 0x1400065f260, {0x14000164360, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x105e868ae, 0x28}, {{0x0, 0x0, 0x0}, 0x1400065f260, {0x14000164360, 0x2c}, {0x1400064dd60, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x105e868ae, 0x28}, {0x105e45865, 0x5}, {{0x0, 0x0, 0x0}, 0x1400065f260, {0x14000164360, 0x2c}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func18.6.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phases", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 72 [sync.Cond.Wait, 9 minutes] sync.runtime_notifyListWait(0x140006e6048, 0x45) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140006e6038) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140006e6030, {0x140006f2000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x14000444b10?}, {0x140006f2000?, 0x1062d9640?, 0x140009818c0?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x1400047b740) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x1400047b740) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x1400044f308) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x1400044f308) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x1400044f308, {0x14000d1ab98, 0xed468, 0x104841c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x14000329088, {0x14000d1ab98, 0xed468, 0xed468}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000d1ab98?, 0x14000eac028?, 0x14000d1aa40?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x1400028dda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140008eb8c0, {0x1341d9600, 0x140008b6900}, {0x1067fd920, 0x14000077320}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 41 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 106 [sync.Cond.Wait, 9 minutes] sync.runtime_notifyListWait(0x14000423cc8, 0x13) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x14000423cb8) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x14000423cb0, {0x14000704000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x14000444b10?}, {0x14000704000?, 0x1062d9640?, 0x14000910840?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x14000076360) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x14000076360) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000898608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000898608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000898608, {0x14000c0d765, 0xfa89b, 0x104841c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400026c2c8, {0x14000c0d765, 0xfa89b, 0xfa89b}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000c0d765?, 0x140009126e8?, 0x14000c0d4cf?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x14000991da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140008eb8c0, {0x1341d9600, 0x140003e1560}, {0x1067fd920, 0x14000077320}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 41 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 152 [syscall, 9 minutes] syscall.syscall6(0x10026ed20?, 0x134011580?, 0x108a553e8?, 0x90?, 0x14000808008?, 0x140008dc1b0?, 0x14000617a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000617a98?, 0x10477e6fc?, 0x90?, 0x10676c660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000260bd0?, 0x14000617ad4, 0x1400026ed20?, 0x14000260b60?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140000ef200) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000e51408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000422300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000422300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a3010?, 0x1068041e8?, 0x14000260930?}}, {0x106824808?, 0x14000260930?}, {0x14000e64000?, 0xc2320beb56221f58?}, {0x1067fd120, 0x140000eef80}, {0x1067fd120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a3010?, 0x107d0b668?, 0x107ce1ee0?}}, {0x106824808, 0x14000260930}, {0x14000e64000, 0xcd}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106828e48?, 0x1400012aba0?}, {0x105e48924?, 0x1400012c5a0?}}, {0x105e868ae, 0x28}, {0x14000184d20, 0x26}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 148 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/09/25 16:10:08.284 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/09/25 16:10:08.284 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/09/25 16:10:08.284 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 @ 10/09/25 16:10:08.284 [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/09/25 16:19:43.188 This is the Progress Report generated when the suite timeout occurred: VirtualMachineLabelAndAnnotation When virtual disks are applied checks VDs phases (Spec Runtime: 9m34.864s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 In [It] (Node Runtime: 9m34.863s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 At [By Step] VDs should be in Ready phases (Step Runtime: 9m34.863s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 Spec Goroutine goroutine 148 [sync.WaitGroup.Wait, 9 minutes] sync.runtime_SemacquireWaitGroup(0x1046fcb30?, 0x60?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x1400026ebe0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140005a0ff0, 0x1, 0x10616ce70?}, {0x105e868ae, 0x28}, {{0x0, 0x0, 0x0}, 0x1400065f260, {0x14000164360, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x105e868ae, 0x28}, {{0x0, 0x0, 0x0}, 0x1400065f260, {0x14000164360, 0x2c}, {0x1400064dd60, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x105e868ae, 0x28}, {0x105e45865, 0x5}, {{0x0, 0x0, 0x0}, 0x1400065f260, {0x14000164360, 0x2c}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func18.6.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phases", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 72 [sync.Cond.Wait, 9 minutes] sync.runtime_notifyListWait(0x140006e6048, 0x45) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140006e6038) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140006e6030, {0x140006f2000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x14000444b10?}, {0x140006f2000?, 0x1062d9640?, 0x140009818c0?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x1400047b740) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x1400047b740) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x1400044f308) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x1400044f308) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x1400044f308, {0x14000d1ab98, 0xed468, 0x104841c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x14000329088, {0x14000d1ab98, 0xed468, 0xed468}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000d1ab98?, 0x14000eac028?, 0x14000d1aa40?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x1400028dda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140008eb8c0, {0x1341d9600, 0x140008b6900}, {0x1067fd920, 0x14000077320}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 41 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 106 [sync.Cond.Wait, 9 minutes] sync.runtime_notifyListWait(0x14000423cc8, 0x13) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x14000423cb8) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x14000423cb0, {0x14000704000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x14000444b10?}, {0x14000704000?, 0x1062d9640?, 0x14000910840?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x14000076360) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x14000076360) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000898608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000898608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000898608, {0x14000c0d765, 0xfa89b, 0x104841c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400026c2c8, {0x14000c0d765, 0xfa89b, 0xfa89b}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000c0d765?, 0x140009126e8?, 0x14000c0d4cf?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x14000991da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140008eb8c0, {0x1341d9600, 0x140003e1560}, {0x1067fd920, 0x14000077320}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 41 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 152 [syscall, 9 minutes] syscall.syscall6(0x10026ed20?, 0x134011580?, 0x108a553e8?, 0x90?, 0x14000808008?, 0x140008dc1b0?, 0x14000617a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000617a98?, 0x10477e6fc?, 0x90?, 0x10676c660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000260bd0?, 0x14000617ad4, 0x1400026ed20?, 0x14000260b60?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140000ef200) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000e51408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000422300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000422300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a3010?, 0x1068041e8?, 0x14000260930?}}, {0x106824808?, 0x14000260930?}, {0x14000e64000?, 0xc2320beb56221f58?}, {0x1067fd120, 0x140000eef80}, {0x1067fd120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a3010?, 0x107d0b668?, 0x107ce1ee0?}}, {0x106824808, 0x14000260930}, {0x14000e64000, 0xcd}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106828e48?, 0x1400012aba0?}, {0x105e48924?, 0x1400012c5a0?}}, {0x105e868ae, 0x28}, {0x14000184d20, 0x26}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 148 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/09/25 16:19:43.194 (9m34.869s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/09/25 16:19:43.194 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/09/25 16:19:47.918 (4.724s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400040b398>: the container "virtualization-controller" was not found: virtualization-controller-789fdbbc5d-ks2cn { errs: [ <*errors.errorString | 0x1400078ccc0>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-789fdbbc5d-ks2cn", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/09/25 16:19:48.173 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 16:19:47.921 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 16:19:47.921 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 16:19:47.921 [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400040b398>: the container "virtualization-controller" was not found: virtualization-controller-789fdbbc5d-ks2cn { errs: [ <*errors.errorString | 0x1400078ccc0>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-789fdbbc5d-ks2cn", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/09/25 16:19:48.173 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 16:19:48.173 (252ms) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/09/25 16:19:48.174 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/09/25 16:19:52.033 (3.859s) + + + \ No newline at end of file diff --git a/artifacts/test-auth-sds-20251009-085635-17171/junit.xml b/artifacts/test-auth-sds-20251009-085635-17171/junit.xml new file mode 100644 index 0000000000..2b5b61fbc9 --- /dev/null +++ b/artifacts/test-auth-sds-20251009-085635-17171/junit.xml @@ -0,0 +1,725 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 09:16:30.037 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 09:16:34.323 (4.286s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 09:16:34.323 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 09:16:34.323 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.328 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:16:34.329 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.329 (1ms) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.329 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:16:34.329 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.329 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.329 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:16:34.33 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.33 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.33 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:16:34.33 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.33 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.33 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:16:34.33 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.33 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.33 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:16:34.33 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.33 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.33 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:16:34.331 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.331 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.331 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:16:34.331 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.331 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.331 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:16:34.331 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.331 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.331 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:16:34.331 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.331 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.331 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:16:34.331 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.331 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.332 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:16:34.332 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:16:34.332 (0s) + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:50 @ 10/09/25 09:16:36.31 + > Enter [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:38 @ 10/09/25 09:16:34.332 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:50 @ 10/09/25 09:16:36.31 < Exit [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:38 @ 10/09/25 09:16:36.311 (1.979s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/09/25 09:16:36.311 < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/09/25 09:16:41.09 (4.779s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:87 @ 10/09/25 09:16:41.091 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:96 @ 10/09/25 09:16:41.091 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:107 @ 10/09/25 09:16:41.091 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:116 @ 10/09/25 09:16:41.091 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:127 @ 10/09/25 09:16:41.091 + + + [FAILED] Warning: resource namespaces/head-45bdb83d-end-to-end-vm-restore-force is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vd.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualdisk?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vd.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualdisk?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vd.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualdisk?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vd.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualdisk?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vi.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualimage?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vi.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualimage?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vm.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/mutate-virtualization-deckhouse-io-v1alpha2-virtualmachine?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vm.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/mutate-virtualization-deckhouse-io-v1alpha2-virtualmachine?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vmbda.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualmachineblockdeviceattachment?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vmbda.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualmachineblockdeviceattachment?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vmbda.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualmachineblockdeviceattachment?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vmbda.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualmachineblockdeviceattachment?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Unexpected error: <*exec.ExitError | 0x140006e12e0>: exit status 1 { ProcessState: { pid: 74588, status: 256, rusage: { Utime: { Sec: 0, Usec: 413506, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 85909, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 72466432, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5230, Majflt: 7, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 122, Msgrcv: 171, Nsignals: 454, Nvcsw: 377, Nivcsw: 5255, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:108 @ 10/09/25 09:16:49.937 + > Enter [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/09/25 09:16:41.091 < Exit [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/09/25 09:16:43.081 (1.99s) > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/09/25 09:16:43.081 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/09/25 09:16:43.081 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/09/25 09:16:43.081 [FAILED] Warning: resource namespaces/head-45bdb83d-end-to-end-vm-restore-force is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vd.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualdisk?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vd.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualdisk?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vd.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualdisk?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vd.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualdisk?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vi.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualimage?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vi.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualimage?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vm.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/mutate-virtualization-deckhouse-io-v1alpha2-virtualmachine?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vm.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/mutate-virtualization-deckhouse-io-v1alpha2-virtualmachine?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vmbda.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualmachineblockdeviceattachment?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vmbda.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualmachineblockdeviceattachment?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vmbda.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualmachineblockdeviceattachment?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-restore-force": Internal error occurred: failed calling webhook "vmbda.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualmachineblockdeviceattachment?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Unexpected error: <*exec.ExitError | 0x140006e12e0>: exit status 1 { ProcessState: { pid: 74588, status: 256, rusage: { Utime: { Sec: 0, Usec: 413506, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 85909, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 72466432, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5230, Majflt: 7, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 122, Msgrcv: 171, Nsignals: 454, Nvcsw: 377, Nivcsw: 5255, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:108 @ 10/09/25 09:16:49.937 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/09/25 09:16:49.937 (6.856s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/09/25 09:16:49.937 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/09/25 09:16:54.618 (4.681s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/09/25 09:16:54.618 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:131 @ 10/09/25 09:16:54.618 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:177 @ 10/09/25 09:16:54.618 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:341 @ 10/09/25 09:16:54.618 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:367 @ 10/09/25 09:16:54.618 + + + + > Enter [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/09/25 09:16:54.619 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:51 @ 10/09/25 09:16:54.619 < Exit [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/09/25 09:16:54.619 (0s) > Enter [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/09/25 09:16:54.619 < Exit [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/09/25 09:16:54.619 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:82 @ 10/09/25 09:16:54.619 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:94 @ 10/09/25 09:16:54.619 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:106 @ 10/09/25 09:16:54.619 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:116 @ 10/09/25 09:16:54.619 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:132 @ 10/09/25 09:16:54.619 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:151 @ 10/09/25 09:16:54.619 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:157 @ 10/09/25 09:16:54.619 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:189 @ 10/09/25 09:16:54.619 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:216 @ 10/09/25 09:16:54.62 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:252 @ 10/09/25 09:16:54.62 + + + + > Enter [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/09/25 09:16:54.62 [SKIPPED] Module SDN is disabled. Skipping all tests for module SDN. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:50 @ 10/09/25 09:16:54.735 < Exit [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/09/25 09:16:54.736 (116ms) > Enter [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/09/25 09:16:54.736 < Exit [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/09/25 09:16:54.736 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/09/25 09:16:54.736 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/09/25 09:16:54.736 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:111 @ 10/09/25 09:16:54.736 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:125 @ 10/09/25 09:16:54.736 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:141 @ 10/09/25 09:16:54.737 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:156 @ 10/09/25 09:16:54.737 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:169 @ 10/09/25 09:16:54.737 + + + [FAILED] Warning: resource namespaces/head-45bdb83d-end-to-end-vm-disk-attachment is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (InternalError): error when creating "/tmp/testdata/vm-disk-attachment": Internal error occurred: failed calling webhook "vd.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualdisk?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-disk-attachment": Internal error occurred: failed calling webhook "vd.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualdisk?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-disk-attachment": Internal error occurred: failed calling webhook "vi.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualimage?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-disk-attachment": Internal error occurred: failed calling webhook "vm.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/mutate-virtualization-deckhouse-io-v1alpha2-virtualmachine?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Expected <bool>: false to equal <bool>: true In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:82 @ 10/09/25 09:16:59.767 + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/09/25 09:16:54.737 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/09/25 09:16:56.386 (1.649s) > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/09/25 09:16:56.386 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/09/25 09:16:56.386 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/09/25 09:16:56.386 [FAILED] Warning: resource namespaces/head-45bdb83d-end-to-end-vm-disk-attachment is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (InternalError): error when creating "/tmp/testdata/vm-disk-attachment": Internal error occurred: failed calling webhook "vd.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualdisk?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-disk-attachment": Internal error occurred: failed calling webhook "vd.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualdisk?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-disk-attachment": Internal error occurred: failed calling webhook "vi.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/validate-virtualization-deckhouse-io-v1alpha2-virtualimage?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Error from server (InternalError): error when creating "/tmp/testdata/vm-disk-attachment": Internal error occurred: failed calling webhook "vm.virtualization-controller.validate.d8-virtualization": failed to call webhook: Post "https://virtualization-controller.d8-virtualization.svc:443/mutate-virtualization-deckhouse-io-v1alpha2-virtualmachine?timeout=10s": dial tcp 10.223.76.249:443: connect: operation not permitted Expected <bool>: false to equal <bool>: true In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:82 @ 10/09/25 09:16:59.767 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/09/25 09:16:59.767 (3.381s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/09/25 09:16:59.767 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/09/25 09:17:04.431 (4.664s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/09/25 09:17:04.431 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/09/25 09:17:04.431 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/09/25 09:17:04.431 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/09/25 09:17:04.432 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/09/25 09:17:04.432 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/09/25 09:17:04.432 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/09/25 09:17:04.432 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/09/25 09:17:04.432 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/09/25 09:17:04.432 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/09/25 09:17:04.432 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/09/25 09:17:04.432 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/09/25 09:17:04.432 + + + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/09/25 09:17:04.432 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/09/25 09:17:06.39 (1.958s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/09/25 09:17:06.391 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/09/25 09:17:10.87 (4.48s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/09/25 09:17:10.87 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/09/25 09:17:10.87 (0s) + + + [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/09/25 09:17:23.592 This is the Progress Report generated when the suite timeout occurred: VirtualMachineConnectivity When virtual images are applied checks VIs phases (Spec Runtime: 12.721s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 In [It] (Node Runtime: 12.721s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 At [By Step] VIs should be in Ready phases (Step Runtime: 12.721s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:97 Spec Goroutine goroutine 195 [sync.WaitGroup.Wait] sync.runtime_SemacquireWaitGroup(0x104134b30?, 0x70?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000efc440) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000039850, 0x1, 0x105ba4e70?}, {0x1058c0d43, 0x29}, {{0x0, 0x0, 0x0}, 0x14000701b90, {0x14000c83bc0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1058c0d43, 0x29}, {{0x0, 0x0, 0x0}, 0x14000701b90, {0x14000c83bc0, 0x25}, {0x140009a8120, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1058c0d43, 0x29}, {0x10587d865, 0x5}, {{0x0, 0x0, 0x0}, 0x14000701b90, {0x14000c83bc0, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.4.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:98 | It("checks VIs phases", func() { | By(fmt.Sprintf("VIs should be in %s phases", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVI, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 9 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 79 [syscall] syscall.syscall6(0x100a09bb0?, 0x151d6dea0?, 0x10cd645c0?, 0x90?, 0x14000600008?, 0x14000fa2a20?, 0x140000b3a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140000b3a98?, 0x1041b66fc?, 0x90?, 0x1061a4660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002d6e00?, 0x140000b3ad4, 0x14000a09bb0?, 0x140002d6d90?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000529500) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000686808?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000018480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000018480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140003f39a0?, 0x10623c1e8?, 0x140002d6cb0?}}, {0x10625c808?, 0x140002d6cb0?}, {0x14000062270?, 0xc231f3b9134ab498?}, {0x106235120, 0x14000529400}, {0x106235120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140003f39a0?, 0x107743668?, 0x107719ee0?}}, {0x10625c808, 0x140002d6cb0}, {0x14000062270, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106260e48?, 0x14000429908?}, {0x105880924?, 0x14000ea46d8?}}, {0x1058c0d43, 0x29}, {0x14000710780, 0x23}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 195 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/09/25 09:17:10.871 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:97 @ 10/09/25 09:17:10.871 [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/09/25 09:17:23.592 This is the Progress Report generated when the suite timeout occurred: VirtualMachineConnectivity When virtual images are applied checks VIs phases (Spec Runtime: 12.721s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 In [It] (Node Runtime: 12.721s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 At [By Step] VIs should be in Ready phases (Step Runtime: 12.721s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:97 Spec Goroutine goroutine 195 [sync.WaitGroup.Wait] sync.runtime_SemacquireWaitGroup(0x104134b30?, 0x70?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000efc440) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000039850, 0x1, 0x105ba4e70?}, {0x1058c0d43, 0x29}, {{0x0, 0x0, 0x0}, 0x14000701b90, {0x14000c83bc0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1058c0d43, 0x29}, {{0x0, 0x0, 0x0}, 0x14000701b90, {0x14000c83bc0, 0x25}, {0x140009a8120, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1058c0d43, 0x29}, {0x10587d865, 0x5}, {{0x0, 0x0, 0x0}, 0x14000701b90, {0x14000c83bc0, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.4.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:98 | It("checks VIs phases", func() { | By(fmt.Sprintf("VIs should be in %s phases", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVI, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 9 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 79 [syscall] syscall.syscall6(0x100a09bb0?, 0x151d6dea0?, 0x10cd645c0?, 0x90?, 0x14000600008?, 0x14000fa2a20?, 0x140000b3a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140000b3a98?, 0x1041b66fc?, 0x90?, 0x1061a4660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002d6e00?, 0x140000b3ad4, 0x14000a09bb0?, 0x140002d6d90?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000529500) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000686808?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000018480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000018480) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140003f39a0?, 0x10623c1e8?, 0x140002d6cb0?}}, {0x10625c808?, 0x140002d6cb0?}, {0x14000062270?, 0xc231f3b9134ab498?}, {0x106235120, 0x14000529400}, {0x106235120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140003f39a0?, 0x107743668?, 0x107719ee0?}}, {0x10625c808, 0x140002d6cb0}, {0x14000062270, 0xc4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106260e48?, 0x14000429908?}, {0x105880924?, 0x14000ea46d8?}}, {0x1058c0d43, 0x29}, {0x14000710780, 0x23}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 195 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/09/25 09:17:23.597 (12.726s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/09/25 09:17:23.597 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/09/25 09:17:28.253 (4.656s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400051e468>: the container "virtualization-controller" was restarted: virtualization-controller-597f769488-l2l4f the container "virtualization-controller" was restarted: virtualization-controller-597f769488-sxvgr the container "virtualization-controller" was not found: virtualization-controller-7c5cb4798d-bc8sf { errs: [ <*errors.joinError | 0x1400051e450>{ errs: [ <*errors.joinError | 0x1400051e438>{ errs: [ <*errors.errorString | 0x1400054c4c0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-597f769488-l2l4f", }, ], }, <*errors.errorString | 0x1400054c530>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-597f769488-sxvgr", }, ], }, <*errors.errorString | 0x1400054c550>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-7c5cb4798d-bc8sf", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/09/25 09:17:28.628 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 09:17:28.254 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 09:17:28.254 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 09:17:28.254 [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400051e468>: the container "virtualization-controller" was restarted: virtualization-controller-597f769488-l2l4f the container "virtualization-controller" was restarted: virtualization-controller-597f769488-sxvgr the container "virtualization-controller" was not found: virtualization-controller-7c5cb4798d-bc8sf { errs: [ <*errors.joinError | 0x1400051e450>{ errs: [ <*errors.joinError | 0x1400051e438>{ errs: [ <*errors.errorString | 0x1400054c4c0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-597f769488-l2l4f", }, ], }, <*errors.errorString | 0x1400054c530>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-597f769488-sxvgr", }, ], }, <*errors.errorString | 0x1400054c550>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-7c5cb4798d-bc8sf", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/09/25 09:17:28.628 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 09:17:28.628 (374ms) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/09/25 09:17:28.628 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/09/25 09:17:32.786 (4.157s) + + + \ No newline at end of file diff --git a/artifacts/test-cephrbd-20251010-084801-21819/junit.xml b/artifacts/test-cephrbd-20251010-084801-21819/junit.xml new file mode 100644 index 0000000000..d8a633a96f --- /dev/null +++ b/artifacts/test-cephrbd-20251010-084801-21819/junit.xml @@ -0,0 +1,2 @@ + + \ No newline at end of file diff --git a/artifacts/test-cleanup-sds-20251009-084413-6548/junit.xml b/artifacts/test-cleanup-sds-20251009-084413-6548/junit.xml new file mode 100644 index 0000000000..18b52e912c --- /dev/null +++ b/artifacts/test-cleanup-sds-20251009-084413-6548/junit.xml @@ -0,0 +1,696 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 09:05:20.13 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 09:05:25.125 (4.996s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 09:05:25.125 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 09:05:25.125 (0s) + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:50 @ 10/09/25 09:05:27.134 + > Enter [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:38 @ 10/09/25 09:05:25.128 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:50 @ 10/09/25 09:05:27.134 < Exit [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:38 @ 10/09/25 09:05:27.134 (2.007s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/09/25 09:05:27.134 < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/09/25 09:05:31.834 (4.7s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:87 @ 10/09/25 09:05:31.834 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:96 @ 10/09/25 09:05:31.835 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:107 @ 10/09/25 09:05:31.835 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:116 @ 10/09/25 09:05:31.835 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:127 @ 10/09/25 09:05:31.835 + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.835 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/09/25 09:05:31.835 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.835 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.836 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/09/25 09:05:31.836 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.836 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.836 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/09/25 09:05:31.836 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.836 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.836 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/09/25 09:05:31.836 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.836 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.836 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/09/25 09:05:31.836 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.836 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.836 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/09/25 09:05:31.837 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.837 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.837 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/09/25 09:05:31.837 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.837 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.837 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/09/25 09:05:31.837 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.837 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.837 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/09/25 09:05:31.837 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/09/25 09:05:31.837 (0s) + + + > Enter [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/09/25 09:05:31.837 < Exit [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/09/25 09:05:33.853 (2.015s) > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/09/25 09:05:33.853 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/09/25 09:05:33.853 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/09/25 09:05:33.853 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/09/25 09:05:36.305 (2.452s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/09/25 09:05:36.305 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/09/25 09:05:36.305 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/09/25 09:05:36.308 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/09/25 09:05:36.308 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/09/25 09:05:36.308 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:68 @ 10/09/25 09:05:36.308 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/09/25 09:05:56.629 (20.321s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/09/25 09:05:56.629 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/09/25 09:05:56.629 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/09/25 09:05:56.63 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/09/25 09:05:56.63 (0s) > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/09/25 09:05:56.63 STEP: VM should be in Running phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:79 @ 10/09/25 09:05:56.63 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/09/25 09:05:59.447 (2.817s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/09/25 09:05:59.447 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/09/25 09:05:59.447 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/09/25 09:05:59.448 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/09/25 09:05:59.448 (0s) > Enter [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/09/25 09:05:59.448 < Exit [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/09/25 09:05:59.448 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/09/25 09:05:59.448 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/09/25 09:05:59.448 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/09/25 09:05:59.448 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/09/25 09:05:59.448 (0s) > Enter [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/09/25 09:05:59.448 < Exit [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/09/25 09:05:59.448 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/09/25 09:05:59.448 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/09/25 09:05:59.448 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/09/25 09:05:59.448 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/09/25 09:05:59.448 (0s) > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/09/25 09:05:59.448 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/09/25 09:05:59.448 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/09/25 09:07:09.925 (1m10.478s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/09/25 09:07:09.926 (1m10.478s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/09/25 09:07:09.926 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/09/25 09:07:09.926 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.926 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:07:09.926 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.926 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.927 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:07:09.927 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.927 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.927 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:07:09.927 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.927 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.927 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:07:09.927 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.927 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.928 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:07:09.928 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.928 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.928 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:07:09.928 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.928 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.928 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:07:09.928 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.928 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.928 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:07:09.928 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.928 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.928 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:07:09.928 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.928 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.929 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:07:09.929 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.929 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.929 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:07:09.929 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.929 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.929 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/09/25 09:07:09.929 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/09/25 09:07:09.929 (0s) + + + > Enter [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/09/25 09:07:09.929 < Exit [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/09/25 09:07:12.06 (2.13s) > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/09/25 09:07:12.06 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/09/25 09:07:12.06 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/09/25 09:07:12.06 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/09/25 09:07:18.975 (6.915s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/09/25 09:07:18.975 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/09/25 09:07:18.975 (0s) + + + [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/09/25 09:10:12.892 This is the Progress Report generated when the suite timeout occurred: VirtualMachineRestoreForce When the virtualization resources are applied checks the resources phase (Spec Runtime: 2m53.891s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 In [It] (Node Runtime: 2m53.891s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 At [By Step] `VirtualMachine` agent should be ready (Step Runtime: 2m53.891s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 Spec Goroutine goroutine 220 [sync.WaitGroup.Wait, 4 minutes] sync.runtime_SemacquireWaitGroup(0x100bfcb30?, 0x20?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x140008bf1d0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x1400081e460, 0x2, 0x10266ce70?}, {0x10238b0d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400053e2a0, {0x1400039fb30, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x10238b0d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400053e2a0, {0x1400039fb30, 0x29}, {0x14000962090, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x10238b0d7, 0x2a}, {0x10234837b, 0x7}, {{0x0, 0x0, 0x0}, 0x1400053e2a0, {0x1400039fb30, 0x29}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x1400053e2a0, {0x1400039fb30, 0x29}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func21.4.2.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 | It("checks the resources phase", func() { | By("`VirtualMachine` agent should be ready", func() { > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: namespace, github.com/onsi/ginkgo/v2/internal.(*Suite).By(0x1400044b508, {0x102382434, 0x26}, {0x1400015ff48, 0x1, 0x1022db584?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:323 github.com/onsi/ginkgo/v2.By({0x102382434?, 0x102d04210?}, {0x14000fc3f48?, 0x140004bcfa0?, 0x14000fc3f58?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:600 > github.com/deckhouse/virtualization/tests/e2e.init.func21.4.2() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 | | It("checks the resources phase", func() { > By("`VirtualMachine` agent should be ready", func() { | WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 7 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 95 [sync.Cond.Wait, 4 minutes] sync.runtime_notifyListWait(0x140008d81c8, 0x27) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140008d81b8) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140008d81b0, {0x140008cd000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x140005bc570?}, {0x140008cd000?, 0x1027d9640?, 0x14000828e70?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140007c8600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140007c8600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x140008ea008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x140008ea008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x140008ea008, {0x14000c08ae0, 0xf7520, 0x100d41c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400029a008, {0x14000c08ae0, 0xf7520, 0xf7520}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000c08ae0?, 0x140002bba98?, 0x14000c08856?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x140007dbda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x14000536450, {0x1105901b0, 0x140009f9980}, {0x102cfd920, 0x14000119380}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 27 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 224 [syscall, 4 minutes] syscall.syscall6(0x100a00e60?, 0x110864660?, 0x104f545c0?, 0x90?, 0x14000680008?, 0x140008ae3f0?, 0x14001021a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14001021a98?, 0x100c7e6fc?, 0x90?, 0x102c6c660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400028a7e0?, 0x14001021ad4, 0x14000a00e60?, 0x1400028a770?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000480700) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x1400097b008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000626180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000626180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002b99a0?, 0x102d041e8?, 0x1400028a690?}}, {0x102d24808?, 0x1400028a690?}, {0x14000064340?, 0xc231f3250106dcf8?}, {0x102cfd120, 0x14000480680}, {0x102cfd120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002b99a0?, 0x10420b668?, 0x1041e1ee0?}}, {0x102d24808, 0x1400028a690}, {0x14000064340, 0xd0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102d28e48?, 0x1400048c210?}, {0x102348924?, 0x14000fc3ed8?}}, {0x10238b0d7, 0x2a}, {0x14000fa36d0, 0x28}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 220 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 225 [syscall, 4 minutes] syscall.syscall6(0x1008bf210?, 0x11060c648?, 0x104f54a78?, 0x90?, 0x140000a5808?, 0x14000868ea0?, 0x1400101fa68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400101fa98?, 0x100c7e6fc?, 0x90?, 0x102c6c660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000280850?, 0x1400101fad4, 0x140008bf210?, 0x140002807e0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140006d2d40) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x1400095f808?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140008d8780) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140008d8780) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002b99a0?, 0x102d041e8?, 0x14000280700?}}, {0x102d24808?, 0x14000280700?}, {0x140008baea0?, 0xc231f3250106b5e8?}, {0x102cfd120, 0x140006d2cc0}, {0x102cfd120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002b99a0?, 0x10420b668?, 0x1041e1ee0?}}, {0x102d24808, 0x14000280700}, {0x140008baea0, 0xc6}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102d28e48?, 0x1400048c210?}, {0x102348924?, 0x0?}}, {0x10238b0d7, 0x2a}, {0x14000fa36f9, 0x1e}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 220 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) + > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/09/25 09:07:18.975 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/09/25 09:07:18.975 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/09/25 09:07:18.975 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/09/25 09:07:18.975 [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/09/25 09:10:12.892 This is the Progress Report generated when the suite timeout occurred: VirtualMachineRestoreForce When the virtualization resources are applied checks the resources phase (Spec Runtime: 2m53.891s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 In [It] (Node Runtime: 2m53.891s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 At [By Step] `VirtualMachine` agent should be ready (Step Runtime: 2m53.891s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 Spec Goroutine goroutine 220 [sync.WaitGroup.Wait, 4 minutes] sync.runtime_SemacquireWaitGroup(0x100bfcb30?, 0x20?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x140008bf1d0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x1400081e460, 0x2, 0x10266ce70?}, {0x10238b0d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400053e2a0, {0x1400039fb30, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x10238b0d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400053e2a0, {0x1400039fb30, 0x29}, {0x14000962090, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x10238b0d7, 0x2a}, {0x10234837b, 0x7}, {{0x0, 0x0, 0x0}, 0x1400053e2a0, {0x1400039fb30, 0x29}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x1400053e2a0, {0x1400039fb30, 0x29}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func21.4.2.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 | It("checks the resources phase", func() { | By("`VirtualMachine` agent should be ready", func() { > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: namespace, github.com/onsi/ginkgo/v2/internal.(*Suite).By(0x1400044b508, {0x102382434, 0x26}, {0x1400015ff48, 0x1, 0x1022db584?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:323 github.com/onsi/ginkgo/v2.By({0x102382434?, 0x102d04210?}, {0x14000fc3f48?, 0x140004bcfa0?, 0x14000fc3f58?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:600 > github.com/deckhouse/virtualization/tests/e2e.init.func21.4.2() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 | | It("checks the resources phase", func() { > By("`VirtualMachine` agent should be ready", func() { | WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 7 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 95 [sync.Cond.Wait, 4 minutes] sync.runtime_notifyListWait(0x140008d81c8, 0x27) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140008d81b8) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140008d81b0, {0x140008cd000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x140005bc570?}, {0x140008cd000?, 0x1027d9640?, 0x14000828e70?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140007c8600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140007c8600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x140008ea008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x140008ea008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x140008ea008, {0x14000c08ae0, 0xf7520, 0x100d41c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400029a008, {0x14000c08ae0, 0xf7520, 0xf7520}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000c08ae0?, 0x140002bba98?, 0x14000c08856?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x140007dbda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x14000536450, {0x1105901b0, 0x140009f9980}, {0x102cfd920, 0x14000119380}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 27 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 224 [syscall, 4 minutes] syscall.syscall6(0x100a00e60?, 0x110864660?, 0x104f545c0?, 0x90?, 0x14000680008?, 0x140008ae3f0?, 0x14001021a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14001021a98?, 0x100c7e6fc?, 0x90?, 0x102c6c660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400028a7e0?, 0x14001021ad4, 0x14000a00e60?, 0x1400028a770?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000480700) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x1400097b008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000626180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000626180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002b99a0?, 0x102d041e8?, 0x1400028a690?}}, {0x102d24808?, 0x1400028a690?}, {0x14000064340?, 0xc231f3250106dcf8?}, {0x102cfd120, 0x14000480680}, {0x102cfd120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002b99a0?, 0x10420b668?, 0x1041e1ee0?}}, {0x102d24808, 0x1400028a690}, {0x14000064340, 0xd0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102d28e48?, 0x1400048c210?}, {0x102348924?, 0x14000fc3ed8?}}, {0x10238b0d7, 0x2a}, {0x14000fa36d0, 0x28}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 220 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 225 [syscall, 4 minutes] syscall.syscall6(0x1008bf210?, 0x11060c648?, 0x104f54a78?, 0x90?, 0x140000a5808?, 0x14000868ea0?, 0x1400101fa68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400101fa98?, 0x100c7e6fc?, 0x90?, 0x102c6c660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000280850?, 0x1400101fad4, 0x140008bf210?, 0x140002807e0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140006d2d40) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x1400095f808?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140008d8780) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140008d8780) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002b99a0?, 0x102d041e8?, 0x14000280700?}}, {0x102d24808?, 0x14000280700?}, {0x140008baea0?, 0xc231f3250106b5e8?}, {0x102cfd120, 0x140006d2cc0}, {0x102cfd120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002b99a0?, 0x10420b668?, 0x1041e1ee0?}}, {0x102d24808, 0x14000280700}, {0x140008baea0, 0xc6}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102d28e48?, 0x1400048c210?}, {0x102348924?, 0x0?}}, {0x10238b0d7, 0x2a}, {0x14000fa36f9, 0x1e}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 220 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/09/25 09:10:12.895 (2m53.894s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/09/25 09:10:12.895 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/09/25 09:10:17.746 (4.851s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x140008191d0>: the container "virtualization-controller" was restarted: virtualization-controller-7845b5698b-g7jl2 the container "virtualization-controller" was not found: virtualization-controller-7d7c98f8d-czdq9 { errs: [ <*errors.joinError | 0x140008191b8>{ errs: [ <*errors.errorString | 0x1400081a260>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-7845b5698b-g7jl2", }, ], }, <*errors.errorString | 0x1400081a290>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-7d7c98f8d-czdq9", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/09/25 09:10:18.01 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 09:10:17.748 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 09:10:17.748 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 09:10:17.748 [FAILED] Expected success, but got an error: <*errors.joinError | 0x140008191d0>: the container "virtualization-controller" was restarted: virtualization-controller-7845b5698b-g7jl2 the container "virtualization-controller" was not found: virtualization-controller-7d7c98f8d-czdq9 { errs: [ <*errors.joinError | 0x140008191b8>{ errs: [ <*errors.errorString | 0x1400081a260>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-7845b5698b-g7jl2", }, ], }, <*errors.errorString | 0x1400081a290>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-7d7c98f8d-czdq9", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/09/25 09:10:18.01 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 09:10:18.01 (263ms) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/09/25 09:10:18.011 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/09/25 09:10:21.806 (3.795s) + + + \ No newline at end of file diff --git a/artifacts/test-fix-sds-20251009-090132-16563/junit.xml b/artifacts/test-fix-sds-20251009-090132-16563/junit.xml new file mode 100644 index 0000000000..a967800ba3 --- /dev/null +++ b/artifacts/test-fix-sds-20251009-090132-16563/junit.xml @@ -0,0 +1,669 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 09:20:05.868 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 09:20:10.193 (4.325s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 09:20:10.193 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 09:20:10.194 (0s) + + + > Enter [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/09/25 09:20:10.195 < Exit [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/09/25 09:20:12.165 (1.97s) > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/09/25 09:20:12.165 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/09/25 09:20:12.165 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/09/25 09:20:12.165 �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "9d2cd021-2d44-4ce0-b6d3-d1fb6d6bd332", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "9d2cd021-2d44-4ce0-b6d3-d1fb6d6bd332", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "9d2cd021-2d44-4ce0-b6d3-d1fb6d6bd332", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "90e9697a-01f6-464c-984a-5b71de907be8", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "90e9697a-01f6-464c-984a-5b71de907be8", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "90e9697a-01f6-464c-984a-5b71de907be8", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "4461bbe2-aca0-41cc-b127-ae5aebc69fa1", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "4461bbe2-aca0-41cc-b127-ae5aebc69fa1", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "4461bbe2-aca0-41cc-b127-ae5aebc69fa1", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "dcc6658c-e59a-42a8-9c41-44626f7b8fbb", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "dcc6658c-e59a-42a8-9c41-44626f7b8fbb", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "dcc6658c-e59a-42a8-9c41-44626f7b8fbb", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5b0ae1da-2b96-4de4-9bdc-6cebc44842b1", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5b0ae1da-2b96-4de4-9bdc-6cebc44842b1", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5b0ae1da-2b96-4de4-9bdc-6cebc44842b1", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "d8b00353-9d6d-4ffd-855b-bddbce57793f", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "d8b00353-9d6d-4ffd-855b-bddbce57793f", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "d8b00353-9d6d-4ffd-855b-bddbce57793f", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "0a697e7a-9e34-4e9a-81d1-df3651b98137", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "0a697e7a-9e34-4e9a-81d1-df3651b98137", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "0a697e7a-9e34-4e9a-81d1-df3651b98137", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "66723a27-4f49-447f-a0f9-6749b19b6d36", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "66723a27-4f49-447f-a0f9-6749b19b6d36", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "66723a27-4f49-447f-a0f9-6749b19b6d36", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "e805a68b-44bf-407b-bcb1-09f91ab15243", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "e805a68b-44bf-407b-bcb1-09f91ab15243", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "e805a68b-44bf-407b-bcb1-09f91ab15243", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "259ac93a-bc3c-478d-9116-a2c2316c59f4", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "259ac93a-bc3c-478d-9116-a2c2316c59f4", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "259ac93a-bc3c-478d-9116-a2c2316c59f4", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "a6ccf856-9a69-49fb-acf3-25be8ba3b90a", "time": "2025-10-09T06:20:15Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "a6ccf856-9a69-49fb-acf3-25be8ba3b90a", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "a6ccf856-9a69-49fb-acf3-25be8ba3b90a", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5c121396-cb7c-4a4d-b62a-95e0c77b36b4", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5c121396-cb7c-4a4d-b62a-95e0c77b36b4", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5c121396-cb7c-4a4d-b62a-95e0c77b36b4", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "ec493041-df89-4eea-a261-6adc418b10c8", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "ec493041-df89-4eea-a261-6adc418b10c8", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "ec493041-df89-4eea-a261-6adc418b10c8", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "c87ebe6b-7f90-44b8-82c5-431799502144", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "c87ebe6b-7f90-44b8-82c5-431799502144", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "c87ebe6b-7f90-44b8-82c5-431799502144", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "656f7f1a-ae1f-4025-ba57-2b376a20540e", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "656f7f1a-ae1f-4025-ba57-2b376a20540e", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "656f7f1a-ae1f-4025-ba57-2b376a20540e", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "8b799964-41ac-4f85-8fbd-8340a5027c72", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "8b799964-41ac-4f85-8fbd-8340a5027c72", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "8b799964-41ac-4f85-8fbd-8340a5027c72", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "66094f95-8e4a-4cac-b050-39821f166a0f", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "66094f95-8e4a-4cac-b050-39821f166a0f", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "66094f95-8e4a-4cac-b050-39821f166a0f", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "8996cf6c-51d6-4176-b80c-096f184c6cc6", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "8996cf6c-51d6-4176-b80c-096f184c6cc6", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "8996cf6c-51d6-4176-b80c-096f184c6cc6", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "18278a3d-c31b-48c3-8a90-07389d754e2f", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "18278a3d-c31b-48c3-8a90-07389d754e2f", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "18278a3d-c31b-48c3-8a90-07389d754e2f", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "7005bad0-9144-4128-a02a-c7601b55208b", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "7005bad0-9144-4128-a02a-c7601b55208b", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "7005bad0-9144-4128-a02a-c7601b55208b", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "2792fafd-294a-4612-96b1-7704fc0d911a", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "2792fafd-294a-4612-96b1-7704fc0d911a", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "2792fafd-294a-4612-96b1-7704fc0d911a", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "36551206-ff82-48e6-865b-1fc21d841113", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "36551206-ff82-48e6-865b-1fc21d841113", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "36551206-ff82-48e6-865b-1fc21d841113", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "110429e5-c649-4b66-925e-464d6ba4de0d", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "110429e5-c649-4b66-925e-464d6ba4de0d", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "110429e5-c649-4b66-925e-464d6ba4de0d", "time": "2025-10-09T06:20:16Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "da7abd21-12c1-4fb9-88a0-f849ecead2ba", "time": "2025-10-09T06:20:17Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "da7abd21-12c1-4fb9-88a0-f849ecead2ba", "time": "2025-10-09T06:20:17Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "da7abd21-12c1-4fb9-88a0-f849ecead2ba", "time": "2025-10-09T06:20:17Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "9ef100b3-54d1-4903-b8da-ad588e6cb9f4", "time": "2025-10-09T06:20:17Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "9ef100b3-54d1-4903-b8da-ad588e6cb9f4", "time": "2025-10-09T06:20:17Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "9ef100b3-54d1-4903-b8da-ad588e6cb9f4", "time": "2025-10-09T06:20:17Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "d3c3ccab-ef12-427f-82cf-337dbd13a85e", "time": "2025-10-09T06:20:17Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "d3c3ccab-ef12-427f-82cf-337dbd13a85e", "time": "2025-10-09T06:20:17Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "d3c3ccab-ef12-427f-82cf-337dbd13a85e", "time": "2025-10-09T06:20:17Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "746eb8a5-9afc-4e54-a263-c5938330001f", "time": "2025-10-09T06:20:17Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "746eb8a5-9afc-4e54-a263-c5938330001f", "time": "2025-10-09T06:20:17Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "746eb8a5-9afc-4e54-a263-c5938330001f", "time": "2025-10-09T06:20:17Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "7d075fe2-fbdd-4f52-8d08-0a6993da4c6e", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "7d075fe2-fbdd-4f52-8d08-0a6993da4c6e", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "7d075fe2-fbdd-4f52-8d08-0a6993da4c6e", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "029064b7-b3ff-4e2c-a903-fcc5f791a94f", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "029064b7-b3ff-4e2c-a903-fcc5f791a94f", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "029064b7-b3ff-4e2c-a903-fcc5f791a94f", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "3388c4ab-aaf3-4391-91ab-f7ee5bdfa0c0", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "3388c4ab-aaf3-4391-91ab-f7ee5bdfa0c0", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "3388c4ab-aaf3-4391-91ab-f7ee5bdfa0c0", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "4a81d209-2120-4701-8c24-05fd46d7de1d", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "4a81d209-2120-4701-8c24-05fd46d7de1d", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "4a81d209-2120-4701-8c24-05fd46d7de1d", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "a03f3b74-7945-42f6-9f53-f6fba50672b1", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "a03f3b74-7945-42f6-9f53-f6fba50672b1", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "a03f3b74-7945-42f6-9f53-f6fba50672b1", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "2a44366f-d1b0-411c-960a-2494b5bf6b0e", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "2a44366f-d1b0-411c-960a-2494b5bf6b0e", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "2a44366f-d1b0-411c-960a-2494b5bf6b0e", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "ded2fd16-0f0a-4bf4-bb53-ff5970a21430", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "ded2fd16-0f0a-4bf4-bb53-ff5970a21430", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "ded2fd16-0f0a-4bf4-bb53-ff5970a21430", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "f0e159b0-396a-461f-b431-9fdd07742ae1", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "f0e159b0-396a-461f-b431-9fdd07742ae1", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "f0e159b0-396a-461f-b431-9fdd07742ae1", "time": "2025-10-09T06:20:18Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "33c82a6a-6ae4-49cb-a92d-e604fd2327f4", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "33c82a6a-6ae4-49cb-a92d-e604fd2327f4", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "33c82a6a-6ae4-49cb-a92d-e604fd2327f4", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "b8f6d354-5a84-4955-bdc9-8a614d6ddb2e", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "b8f6d354-5a84-4955-bdc9-8a614d6ddb2e", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "b8f6d354-5a84-4955-bdc9-8a614d6ddb2e", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "27169726-3f33-41ea-aab1-56739827feef", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "27169726-3f33-41ea-aab1-56739827feef", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "27169726-3f33-41ea-aab1-56739827feef", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "01b8a99d-f226-4dad-a055-c2b3eb89bd4f", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "01b8a99d-f226-4dad-a055-c2b3eb89bd4f", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "01b8a99d-f226-4dad-a055-c2b3eb89bd4f", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "96ddb7c1-d40f-45ea-82b7-06dcd0936c88", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "96ddb7c1-d40f-45ea-82b7-06dcd0936c88", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "96ddb7c1-d40f-45ea-82b7-06dcd0936c88", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "adf286b8-2ce7-4e1d-8bfa-1cccd470ed34", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "adf286b8-2ce7-4e1d-8bfa-1cccd470ed34", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "adf286b8-2ce7-4e1d-8bfa-1cccd470ed34", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "963e4333-0c50-4b9b-beca-53021d4b18c4", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "963e4333-0c50-4b9b-beca-53021d4b18c4", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "963e4333-0c50-4b9b-beca-53021d4b18c4", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5d82290a-6468-4047-859e-b3c3f1b9d00b", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5d82290a-6468-4047-859e-b3c3f1b9d00b", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5d82290a-6468-4047-859e-b3c3f1b9d00b", "time": "2025-10-09T06:20:19Z" }�[0m < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/09/25 09:20:19.38 (7.215s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/09/25 09:20:19.38 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/09/25 09:20:19.38 (0s) + + + [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/09/25 09:21:00.772 This is the Progress Report generated when the suite timeout occurred: SizingPolicy When virtual images are applied checks VIs phases (Spec Runtime: 41.391s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 In [It] (Node Runtime: 41.391s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 At [By Step] VIs should be in Ready phases (Step Runtime: 41.391s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:91 Spec Goroutine goroutine 56 [sync.WaitGroup.Wait] sync.runtime_SemacquireWaitGroup(0x104a78b30?, 0x80?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000a84490) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140006c4400, 0x1, 0x1064e8e70?}, {0x106204d43, 0x29}, {{0x0, 0x0, 0x0}, 0x14000553620, {0x1400048cb40, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x106204d43, 0x29}, {{0x0, 0x0, 0x0}, 0x14000553620, {0x1400048cb40, 0x26}, {0x140007e4160, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x106204d43, 0x29}, {0x1061c1865, 0x5}, {{0x0, 0x0, 0x0}, 0x14000553620, {0x1400048cb40, 0x26}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func9.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:92 | It("checks VIs phases", func() { | By(fmt.Sprintf("VIs should be in %s phases", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVI, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x14000514180?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 119 [sync.Cond.Wait] sync.runtime_notifyListWait(0x140006c27c8, 0x228) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140006c27b8) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140006c27b0, {0x140009b0000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400052a150?}, {0x140009b0000?, 0x106655640?, 0x140008d4dc0?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x14000118720) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x14000118720) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x140009b2008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x140009b2008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x140009b2008, {0x14000b03216, 0x9edea, 0x104bbdc74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400027c008, {0x14000b03216, 0x9edea, 0x9edea}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000b03216?, 0x140008d3588?, 0x14000b02f68?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x14000733da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140004bc870, {0x11461fd58, 0x140008fa090}, {0x106b79920, 0x14000075320}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 15 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 141 [syscall] syscall.syscall6(0x100a844d0?, 0x1090a6448?, 0x108dd18a0?, 0x90?, 0x14000500808?, 0x1400021a090?, 0x14000771a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000771a98?, 0x104afa6fc?, 0x90?, 0x106ae8660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029c380?, 0x14000771ad4, 0x14000a844d0?, 0x1400029c310?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000994200) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000f7c408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000171500) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000171500) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a3470?, 0x106b801e8?, 0x1400029c230?}}, {0x106ba0808?, 0x1400029c230?}, {0x14000064240?, 0xc231f3e830cdaae8?}, {0x106b79120, 0x14000994180}, {0x106b79120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a3470?, 0x108087668?, 0x10805dee0?}}, {0x106ba0808, 0x1400029c230}, {0x14000064240, 0xbe}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106ba4e48?, 0x1400000fe48?}, {0x1061c4924?, 0x14000f96680?}}, {0x106204d43, 0x29}, {0x140008260c0, 0x1c}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 56 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/09/25 09:20:19.38 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/09/25 09:20:19.38 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/09/25 09:20:19.38 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:91 @ 10/09/25 09:20:19.38 �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "9648c457-58d3-4436-b86b-bc88e44d2e61", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "9648c457-58d3-4436-b86b-bc88e44d2e61", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "9648c457-58d3-4436-b86b-bc88e44d2e61", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5e373c4f-127a-4250-923b-5532e20d415d", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5e373c4f-127a-4250-923b-5532e20d415d", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5e373c4f-127a-4250-923b-5532e20d415d", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "dd3cd82b-e276-4323-8549-104ed92b92eb", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "dd3cd82b-e276-4323-8549-104ed92b92eb", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "dd3cd82b-e276-4323-8549-104ed92b92eb", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "cbfcb5a9-80a4-4100-804e-95e5de9456d3", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "cbfcb5a9-80a4-4100-804e-95e5de9456d3", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "cbfcb5a9-80a4-4100-804e-95e5de9456d3", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "11dc4050-5753-449b-a746-9c8d210d8982", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "11dc4050-5753-449b-a746-9c8d210d8982", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "11dc4050-5753-449b-a746-9c8d210d8982", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "0f739caf-4dc0-4a6c-828f-016832288322", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "0f739caf-4dc0-4a6c-828f-016832288322", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "0f739caf-4dc0-4a6c-828f-016832288322", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "f596be85-56da-47f3-9e57-7e1a363b9c64", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "f596be85-56da-47f3-9e57-7e1a363b9c64", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "f596be85-56da-47f3-9e57-7e1a363b9c64", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "d435c265-e5ca-4c94-9415-9b9c74ccf46e", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "d435c265-e5ca-4c94-9415-9b9c74ccf46e", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "d435c265-e5ca-4c94-9415-9b9c74ccf46e", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "62d81cfc-febf-4ac3-9f19-c0b0559d8dcc", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "62d81cfc-febf-4ac3-9f19-c0b0559d8dcc", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "62d81cfc-febf-4ac3-9f19-c0b0559d8dcc", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5003585a-c947-493e-a85d-f9d959c85191", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5003585a-c947-493e-a85d-f9d959c85191", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5003585a-c947-493e-a85d-f9d959c85191", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "6b65656b-f398-4a58-8d55-78f1b236a782", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "6b65656b-f398-4a58-8d55-78f1b236a782", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "6b65656b-f398-4a58-8d55-78f1b236a782", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "4c838b36-2fd9-489e-8560-b0c77cb68b67", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "4c838b36-2fd9-489e-8560-b0c77cb68b67", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "4c838b36-2fd9-489e-8560-b0c77cb68b67", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "3f414b6a-b284-4aee-8dbb-cac022e26938", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "3f414b6a-b284-4aee-8dbb-cac022e26938", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "3f414b6a-b284-4aee-8dbb-cac022e26938", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "44c2b112-d071-4c0b-ba76-f53de87f42c2", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "44c2b112-d071-4c0b-ba76-f53de87f42c2", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "44c2b112-d071-4c0b-ba76-f53de87f42c2", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "943ee62c-aef8-41e8-bcad-f776ec03b7a0", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "943ee62c-aef8-41e8-bcad-f776ec03b7a0", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "943ee62c-aef8-41e8-bcad-f776ec03b7a0", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "e9faf560-3375-4644-8a00-b1d4365a91a7", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "e9faf560-3375-4644-8a00-b1d4365a91a7", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "e9faf560-3375-4644-8a00-b1d4365a91a7", "time": "2025-10-09T06:20:19Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "4bb98401-94c3-46e0-9ccf-829d2bedca9c", "time": "2025-10-09T06:20:20Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "4bb98401-94c3-46e0-9ccf-829d2bedca9c", "time": "2025-10-09T06:20:20Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "4bb98401-94c3-46e0-9ccf-829d2bedca9c", "time": "2025-10-09T06:20:20Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "4a1de953-c007-4509-86f9-1858ad25963b", "time": "2025-10-09T06:20:20Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "4a1de953-c007-4509-86f9-1858ad25963b", "time": "2025-10-09T06:20:20Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "4a1de953-c007-4509-86f9-1858ad25963b", "time": "2025-10-09T06:20:20Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "431240f3-c34e-4065-8aa9-99ccea98fb17", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "431240f3-c34e-4065-8aa9-99ccea98fb17", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "431240f3-c34e-4065-8aa9-99ccea98fb17", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "3d179543-b1fe-43f3-99fe-9d11cd4dcd09", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "3d179543-b1fe-43f3-99fe-9d11cd4dcd09", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "3d179543-b1fe-43f3-99fe-9d11cd4dcd09", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "a693e670-9fa9-4f32-9f41-60c0c3b81c6d", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "a693e670-9fa9-4f32-9f41-60c0c3b81c6d", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "a693e670-9fa9-4f32-9f41-60c0c3b81c6d", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "3bf5a1ef-dfc5-4229-a0d5-5220d84573d1", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "3bf5a1ef-dfc5-4229-a0d5-5220d84573d1", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "3bf5a1ef-dfc5-4229-a0d5-5220d84573d1", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "e69e4e8e-e683-4dee-b326-dc2e25ed3f8f", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "e69e4e8e-e683-4dee-b326-dc2e25ed3f8f", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "e69e4e8e-e683-4dee-b326-dc2e25ed3f8f", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "811fe1f6-9ead-44bc-8418-e496916ba92c", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "811fe1f6-9ead-44bc-8418-e496916ba92c", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "811fe1f6-9ead-44bc-8418-e496916ba92c", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "0e7a30dc-6e38-4c1d-b7a6-ad4d7f6efb98", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "0e7a30dc-6e38-4c1d-b7a6-ad4d7f6efb98", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "0e7a30dc-6e38-4c1d-b7a6-ad4d7f6efb98", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "75097761-c663-48fa-a462-831fe3d5b0e3", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "75097761-c663-48fa-a462-831fe3d5b0e3", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "75097761-c663-48fa-a462-831fe3d5b0e3", "time": "2025-10-09T06:20:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "a1e2ee14-b66a-4ccd-abda-898625c981d5", "time": "2025-10-09T06:20:23Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "a1e2ee14-b66a-4ccd-abda-898625c981d5", "time": "2025-10-09T06:20:23Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "a1e2ee14-b66a-4ccd-abda-898625c981d5", "time": "2025-10-09T06:20:23Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "563804d6-7105-4ea4-9f8f-17c11f52dd36", "time": "2025-10-09T06:20:23Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "563804d6-7105-4ea4-9f8f-17c11f52dd36", "time": "2025-10-09T06:20:23Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "563804d6-7105-4ea4-9f8f-17c11f52dd36", "time": "2025-10-09T06:20:23Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "50706e3d-c34f-4b9a-87a5-3915e3623d40", "time": "2025-10-09T06:20:23Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "50706e3d-c34f-4b9a-87a5-3915e3623d40", "time": "2025-10-09T06:20:23Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "50706e3d-c34f-4b9a-87a5-3915e3623d40", "time": "2025-10-09T06:20:23Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "eee23e64-75ff-4215-a62d-7f9793d8cf39", "time": "2025-10-09T06:20:23Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "eee23e64-75ff-4215-a62d-7f9793d8cf39", "time": "2025-10-09T06:20:23Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "eee23e64-75ff-4215-a62d-7f9793d8cf39", "time": "2025-10-09T06:20:23Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "10914226-8f4a-4940-a129-6b731fd4c94f", "time": "2025-10-09T06:20:24Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "10914226-8f4a-4940-a129-6b731fd4c94f", "time": "2025-10-09T06:20:24Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "10914226-8f4a-4940-a129-6b731fd4c94f", "time": "2025-10-09T06:20:24Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "12b53183-8cd7-4586-b8c3-3beb57fab9ac", "time": "2025-10-09T06:20:24Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "12b53183-8cd7-4586-b8c3-3beb57fab9ac", "time": "2025-10-09T06:20:24Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "12b53183-8cd7-4586-b8c3-3beb57fab9ac", "time": "2025-10-09T06:20:24Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "02dd61a0-98e6-41a2-a930-fb31c93afcc7", "time": "2025-10-09T06:20:25Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "02dd61a0-98e6-41a2-a930-fb31c93afcc7", "time": "2025-10-09T06:20:25Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "02dd61a0-98e6-41a2-a930-fb31c93afcc7", "time": "2025-10-09T06:20:25Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "d1fb4f7a-efa5-46ce-8718-72fb680628d7", "time": "2025-10-09T06:20:25Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "d1fb4f7a-efa5-46ce-8718-72fb680628d7", "time": "2025-10-09T06:20:25Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "d1fb4f7a-efa5-46ce-8718-72fb680628d7", "time": "2025-10-09T06:20:25Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "043f0d07-b69a-40a0-b68c-cc1d094a90da", "time": "2025-10-09T06:20:25Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "043f0d07-b69a-40a0-b68c-cc1d094a90da", "time": "2025-10-09T06:20:25Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "043f0d07-b69a-40a0-b68c-cc1d094a90da", "time": "2025-10-09T06:20:25Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "eeab5622-e346-49c8-833f-e8edeb675aa3", "time": "2025-10-09T06:20:25Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "eeab5622-e346-49c8-833f-e8edeb675aa3", "time": "2025-10-09T06:20:25Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "eeab5622-e346-49c8-833f-e8edeb675aa3", "time": "2025-10-09T06:20:25Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "a162a64b-3dc8-4bff-abce-68cd85665f1c", "time": "2025-10-09T06:20:26Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "a162a64b-3dc8-4bff-abce-68cd85665f1c", "time": "2025-10-09T06:20:26Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "a162a64b-3dc8-4bff-abce-68cd85665f1c", "time": "2025-10-09T06:20:26Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5440c3a7-f667-4773-965c-6dd98816d7f0", "time": "2025-10-09T06:20:26Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5440c3a7-f667-4773-965c-6dd98816d7f0", "time": "2025-10-09T06:20:26Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5440c3a7-f667-4773-965c-6dd98816d7f0", "time": "2025-10-09T06:20:26Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "9470f319-0c38-4081-b4cc-380338fe5168", "time": "2025-10-09T06:20:27Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "9470f319-0c38-4081-b4cc-380338fe5168", "time": "2025-10-09T06:20:27Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "9470f319-0c38-4081-b4cc-380338fe5168", "time": "2025-10-09T06:20:27Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "c7b094d2-6fa5-4f16-ab53-749d6cbafb9a", "time": "2025-10-09T06:20:27Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "c7b094d2-6fa5-4f16-ab53-749d6cbafb9a", "time": "2025-10-09T06:20:27Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "c7b094d2-6fa5-4f16-ab53-749d6cbafb9a", "time": "2025-10-09T06:20:27Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5e0e9292-6416-4dca-a97e-1333d200313b", "time": "2025-10-09T06:20:27Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5e0e9292-6416-4dca-a97e-1333d200313b", "time": "2025-10-09T06:20:27Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "5e0e9292-6416-4dca-a97e-1333d200313b", "time": "2025-10-09T06:20:27Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "26a1a0ab-d5c6-4fb5-9ce9-499bddc417c5", "time": "2025-10-09T06:20:27Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "26a1a0ab-d5c6-4fb5-9ce9-499bddc417c5", "time": "2025-10-09T06:20:27Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "26a1a0ab-d5c6-4fb5-9ce9-499bddc417c5", "time": "2025-10-09T06:20:27Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "6431328e-8292-4db2-abcc-3e5571d18cbf", "time": "2025-10-09T06:20:29Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "6431328e-8292-4db2-abcc-3e5571d18cbf", "time": "2025-10-09T06:20:29Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "6431328e-8292-4db2-abcc-3e5571d18cbf", "time": "2025-10-09T06:20:29Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "1f107ac3-256a-49c6-9e8a-8c67308411b0", "time": "2025-10-09T06:20:29Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "1f107ac3-256a-49c6-9e8a-8c67308411b0", "time": "2025-10-09T06:20:29Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-creating", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "1f107ac3-256a-49c6-9e8a-8c67308411b0", "time": "2025-10-09T06:20:29Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "931747a1-1979-46a7-ab1d-744273c98f32", "time": "2025-10-09T06:20:30Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "931747a1-1979-46a7-ab1d-744273c98f32", "time": "2025-10-09T06:20:30Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "931747a1-1979-46a7-ab1d-744273c98f32", "time": "2025-10-09T06:20:30Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "ea2f6371-e139-473e-be16-6fe004ea9cb8", "time": "2025-10-09T06:20:30Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "ea2f6371-e139-473e-be16-6fe004ea9cb8", "time": "2025-10-09T06:20:30Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-existing-vmclass", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "ea2f6371-e139-473e-be16-6fe004ea9cb8", "time": "2025-10-09T06:20:30Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "729a1fc8-d239-4063-89eb-525212acdfd0", "time": "2025-10-09T06:20:30Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "729a1fc8-d239-4063-89eb-525212acdfd0", "time": "2025-10-09T06:20:30Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "729a1fc8-d239-4063-89eb-525212acdfd0", "time": "2025-10-09T06:20:30Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "The handler failed with an error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "LifeCycleHandler", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "42cffbf8-5545-4625-8de5-c0d63b7d2d8e", "time": "2025-10-09T06:20:30Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "42cffbf8-5545-4625-8de5-c0d63b7d2d8e", "time": "2025-10-09T06:20:30Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "failed to sync virtual disk data source blank: get volume and access modes: storage profile \"sds-local-storage\" not found: storage profile not found", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-45bdb83d-vd-blank-not-existing-vmclass-with-changing", "namespace": "head-45bdb83d-end-to-end-sizing-policy", "reconcileID": "42cffbf8-5545-4625-8de5-c0d63b7d2d8e", "time": "2025-10-09T06:20:30Z" }�[0m [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/09/25 09:21:00.772 This is the Progress Report generated when the suite timeout occurred: SizingPolicy When virtual images are applied checks VIs phases (Spec Runtime: 41.391s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 In [It] (Node Runtime: 41.391s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 At [By Step] VIs should be in Ready phases (Step Runtime: 41.391s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:91 Spec Goroutine goroutine 56 [sync.WaitGroup.Wait] sync.runtime_SemacquireWaitGroup(0x104a78b30?, 0x80?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000a84490) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140006c4400, 0x1, 0x1064e8e70?}, {0x106204d43, 0x29}, {{0x0, 0x0, 0x0}, 0x14000553620, {0x1400048cb40, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x106204d43, 0x29}, {{0x0, 0x0, 0x0}, 0x14000553620, {0x1400048cb40, 0x26}, {0x140007e4160, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x106204d43, 0x29}, {0x1061c1865, 0x5}, {{0x0, 0x0, 0x0}, 0x14000553620, {0x1400048cb40, 0x26}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func9.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:92 | It("checks VIs phases", func() { | By(fmt.Sprintf("VIs should be in %s phases", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVI, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x14000514180?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 119 [sync.Cond.Wait] sync.runtime_notifyListWait(0x140006c27c8, 0x228) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140006c27b8) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140006c27b0, {0x140009b0000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400052a150?}, {0x140009b0000?, 0x106655640?, 0x140008d4dc0?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x14000118720) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x14000118720) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x140009b2008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x140009b2008) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x140009b2008, {0x14000b03216, 0x9edea, 0x104bbdc74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400027c008, {0x14000b03216, 0x9edea, 0x9edea}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000b03216?, 0x140008d3588?, 0x14000b02f68?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x14000733da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140004bc870, {0x11461fd58, 0x140008fa090}, {0x106b79920, 0x14000075320}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 15 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 141 [syscall] syscall.syscall6(0x100a844d0?, 0x1090a6448?, 0x108dd18a0?, 0x90?, 0x14000500808?, 0x1400021a090?, 0x14000771a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000771a98?, 0x104afa6fc?, 0x90?, 0x106ae8660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029c380?, 0x14000771ad4, 0x14000a844d0?, 0x1400029c310?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000994200) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000f7c408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000171500) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000171500) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a3470?, 0x106b801e8?, 0x1400029c230?}}, {0x106ba0808?, 0x1400029c230?}, {0x14000064240?, 0xc231f3e830cdaae8?}, {0x106b79120, 0x14000994180}, {0x106b79120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a3470?, 0x108087668?, 0x10805dee0?}}, {0x106ba0808, 0x1400029c230}, {0x14000064240, 0xbe}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106ba4e48?, 0x1400000fe48?}, {0x1061c4924?, 0x14000f96680?}}, {0x106204d43, 0x29}, {0x140008260c0, 0x1c}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 56 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/09/25 09:21:00.775 (41.394s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/09/25 09:21:00.775 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/09/25 09:21:05.879 (5.104s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x14000cf1878>: the container "virtualization-controller" was restarted: virtualization-controller-59bcb5f6db-mqg8n the container "virtualization-controller" was not found: virtualization-controller-6496c794f4-5vflb { errs: [ <*errors.joinError | 0x14000cf1860>{ errs: [ <*errors.errorString | 0x140006e97a0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-59bcb5f6db-mqg8n", }, ], }, <*errors.errorString | 0x140006e97d0>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-6496c794f4-5vflb", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/09/25 09:21:06.141 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 09:21:05.881 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 09:21:05.881 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 09:21:05.881 [FAILED] Expected success, but got an error: <*errors.joinError | 0x14000cf1878>: the container "virtualization-controller" was restarted: virtualization-controller-59bcb5f6db-mqg8n the container "virtualization-controller" was not found: virtualization-controller-6496c794f4-5vflb { errs: [ <*errors.joinError | 0x14000cf1860>{ errs: [ <*errors.errorString | 0x140006e97a0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-59bcb5f6db-mqg8n", }, ], }, <*errors.errorString | 0x140006e97d0>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-6496c794f4-5vflb", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/09/25 09:21:06.141 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 09:21:06.142 (260ms) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/09/25 09:21:06.142 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/09/25 09:21:13.884 (7.742s) + + + \ No newline at end of file diff --git a/artifacts/test-fix3-sds-20251009-085329-23071/junit.xml b/artifacts/test-fix3-sds-20251009-085329-23071/junit.xml new file mode 100644 index 0000000000..a108a3ca95 --- /dev/null +++ b/artifacts/test-fix3-sds-20251009-085329-23071/junit.xml @@ -0,0 +1,682 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 09:13:29.763 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 09:13:34.84 (5.077s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 09:13:34.84 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/09/25 09:13:34.841 (0s) + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:66 @ 10/09/25 09:13:36.813 + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:54 @ 10/09/25 09:13:34.844 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:66 @ 10/09/25 09:13:36.813 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:54 @ 10/09/25 09:13:36.814 (1.97s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/09/25 09:13:36.814 < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/09/25 09:13:41.881 (5.067s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:95 @ 10/09/25 09:13:41.881 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:106 @ 10/09/25 09:13:41.881 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:117 @ 10/09/25 09:13:41.881 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:128 @ 10/09/25 09:13:41.881 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:139 @ 10/09/25 09:13:41.882 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:159 @ 10/09/25 09:13:41.882 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:180 @ 10/09/25 09:13:41.882 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:224 @ 10/09/25 09:13:41.882 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:285 @ 10/09/25 09:13:41.882 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:331 @ 10/09/25 09:13:41.882 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:357 @ 10/09/25 09:13:41.882 + + + + > Enter [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/09/25 09:13:41.882 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:41 @ 10/09/25 09:13:41.882 < Exit [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/09/25 09:13:41.883 (0s) > Enter [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/09/25 09:13:41.883 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/09/25 09:13:41.883 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/09/25 09:13:46.432 (4.549s) < Exit [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/09/25 09:13:46.432 (4.549s) + + + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:51 @ 10/09/25 09:13:46.432 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:51 @ 10/09/25 09:13:48.467 (2.035s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:61 @ 10/09/25 09:13:48.467 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:61 @ 10/09/25 09:14:16.504 (28.037s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 09:14:16.504 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 09:14:16.504 (0s) + + + > Enter [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:82 @ 10/09/25 09:14:16.505 < Exit [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:82 @ 10/09/25 09:14:17.835 (1.331s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 09:14:17.836 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 09:14:17.836 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:97 @ 10/09/25 09:14:17.836 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:98 @ 10/09/25 09:14:17.836 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:97 @ 10/09/25 09:14:18.98 (1.145s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 09:14:18.981 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 09:14:18.981 (0s) + + + > Enter [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:108 @ 10/09/25 09:14:18.981 STEP: CVIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:109 @ 10/09/25 09:14:18.981 < Exit [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:108 @ 10/09/25 09:14:20.358 (1.377s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 09:14:20.358 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 09:14:20.358 (0s) + + + > Enter [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:119 @ 10/09/25 09:14:20.359 STEP: VMClasses should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:120 @ 10/09/25 09:14:20.359 < Exit [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:119 @ 10/09/25 09:14:21.531 (1.172s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 09:14:21.531 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 09:14:21.531 (0s) + + + [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 @ 10/09/25 09:14:22.736 This is the Progress Report generated when the suite timeout occurred: ComplexTest When virtual machines IP addresses are applied patches custom VMIP with unassigned address (Spec Runtime: 1.205s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 In [It] (Node Runtime: 1.205s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 Spec Goroutine goroutine 196 [syscall] syscall.syscall6(0x100570570?, 0x104859110?, 0x10484c5c0?, 0x90?, 0x14000100808?, 0x1400022a630?, 0x140008dee38?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140008dee68?, 0x10016e6fc?, 0x90?, 0x10215c660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002d6bd0?, 0x140008deea4, 0x14000570570?, 0x140002d6b60?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x1400067e840) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000672408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140007ac300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140007ac300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400040d9a0?, 0x1021f41e8?, 0x140002d6a80?}}, {0x102214808?, 0x140002d6a80?}, {0x14001034680?, 0xc231f29228d1d828?}, {0x1021ed120, 0x1400067e7c0}, {0x1021ed120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400040d9a0?, 0x1036fb668?, 0x1036d1ee0?}}, {0x102214808, 0x140002d6a80}, {0x14001034680, 0xc9}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.PatchResource({{0x102218e48?, 0x14000429ea8?}, {0x101838924?, 0x1036fba88?}}, {0x10189ba93, 0x35}, {0x140002bc620, 0x1a}, {{0x14000878690, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:291 | ctx, cancel := context.WithTimeout(context.Background(), ShortTimeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.MergePatchResource({0x10189ba93, 0x35}, {0x14000878690, 0x25}, {0x140002bc620, 0x1a}, {0x1400064afc0, 0x21}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:166 | func MergePatchResource(resource kc.Resource, ns, name, patch string) error { | GinkgoHelper() > res := kubectl.PatchResource(resource, name, kc.PatchOptions{ | Namespace: ns, | MergePatch: patch, > github.com/deckhouse/virtualization/tests/e2e.AssignIPToVMIP(0x14000578240?, {0x14000878690, 0x25}, {0x140002bc620, 0x1a}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:570 | | patch := fmt.Sprintf(`{"spec":{"staticIP":%q}}`, unassignedIP) > err = MergePatchResource(kc.ResourceVMIP, vmipNamespace, vmipName, patch) | if err != nil { | return fmt.Errorf("%s\n%w", assignErr, err) > github.com/deckhouse/virtualization/tests/e2e.init.func11.7.1.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:133 | vmipName := fmt.Sprintf("%s-%s", namePrefix, "vm-custom-ip") | Eventually(func() error { > return AssignIPToVMIP(f, ns, vmipName) | }).WithTimeout(LongWaitDuration).WithPolling(Interval).Should(Succeed()) | }) reflect.Value.call({0x101cd2f00?, 0x140007d15c0?, 0x14000100808?}, {0x101834ea6, 0x4}, {0x1036fb668, 0x0, 0x10011dcac?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/reflect/value.go:581 reflect.Value.Call({0x101cd2f00?, 0x140007d15c0?, 0x140002c0940?}, {0x1036fb668?, 0x101cd2f00?, 0x140007d15c0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/reflect/value.go:365 github.com/onsi/gomega/internal.(*AsyncAssertion).buildActualPoller.func3() /Users/antont/go/pkg/mod/github.com/onsi/gomega@v1.36.1/internal/async_assertion.go:325 github.com/onsi/gomega/internal.(*AsyncAssertion).match(0x140002c0850, {0x1021fecc8, 0x1036fb668}, 0x1, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/gomega@v1.36.1/internal/async_assertion.go:398 github.com/onsi/gomega/internal.(*AsyncAssertion).Should(0x140002c0850, {0x1021fecc8, 0x1036fb668}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/gomega@v1.36.1/internal/async_assertion.go:145 > github.com/deckhouse/virtualization/tests/e2e.init.func11.7.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 | Eventually(func() error { | return AssignIPToVMIP(f, ns, vmipName) > }).WithTimeout(LongWaitDuration).WithPolling(Interval).Should(Succeed()) | }) | github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 116 [sync.Cond.Wait] sync.runtime_notifyListWait(0x1400087edc8, 0x2ff) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x1400087edb8) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x1400087edb0, {0x140009d7000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x140009d0a20?}, {0x140009d7000?, 0x101cc9640?, 0x140008b0e70?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x14000819440) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x14000819440) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x140009a0608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x140009a0608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x140009a0608, {0x14000b6d285, 0x92d7b, 0x100231c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400029a008, {0x14000b6d285, 0x92d7b, 0x92d7b}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000b6d285?, 0x140008b7108?, 0x14000b6d158?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x1400099bda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140009d0480, {0x12fa5dc00, 0x1400090cb40}, {0x1021ed920, 0x14000075860}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 99 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | + > Enter [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 @ 10/09/25 09:14:21.531 [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 @ 10/09/25 09:14:22.736 This is the Progress Report generated when the suite timeout occurred: ComplexTest When virtual machines IP addresses are applied patches custom VMIP with unassigned address (Spec Runtime: 1.205s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 In [It] (Node Runtime: 1.205s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 Spec Goroutine goroutine 196 [syscall] syscall.syscall6(0x100570570?, 0x104859110?, 0x10484c5c0?, 0x90?, 0x14000100808?, 0x1400022a630?, 0x140008dee38?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140008dee68?, 0x10016e6fc?, 0x90?, 0x10215c660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002d6bd0?, 0x140008deea4, 0x14000570570?, 0x140002d6b60?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x1400067e840) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000672408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140007ac300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140007ac300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400040d9a0?, 0x1021f41e8?, 0x140002d6a80?}}, {0x102214808?, 0x140002d6a80?}, {0x14001034680?, 0xc231f29228d1d828?}, {0x1021ed120, 0x1400067e7c0}, {0x1021ed120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400040d9a0?, 0x1036fb668?, 0x1036d1ee0?}}, {0x102214808, 0x140002d6a80}, {0x14001034680, 0xc9}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.PatchResource({{0x102218e48?, 0x14000429ea8?}, {0x101838924?, 0x1036fba88?}}, {0x10189ba93, 0x35}, {0x140002bc620, 0x1a}, {{0x14000878690, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:291 | ctx, cancel := context.WithTimeout(context.Background(), ShortTimeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.MergePatchResource({0x10189ba93, 0x35}, {0x14000878690, 0x25}, {0x140002bc620, 0x1a}, {0x1400064afc0, 0x21}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:166 | func MergePatchResource(resource kc.Resource, ns, name, patch string) error { | GinkgoHelper() > res := kubectl.PatchResource(resource, name, kc.PatchOptions{ | Namespace: ns, | MergePatch: patch, > github.com/deckhouse/virtualization/tests/e2e.AssignIPToVMIP(0x14000578240?, {0x14000878690, 0x25}, {0x140002bc620, 0x1a}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:570 | | patch := fmt.Sprintf(`{"spec":{"staticIP":%q}}`, unassignedIP) > err = MergePatchResource(kc.ResourceVMIP, vmipNamespace, vmipName, patch) | if err != nil { | return fmt.Errorf("%s\n%w", assignErr, err) > github.com/deckhouse/virtualization/tests/e2e.init.func11.7.1.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:133 | vmipName := fmt.Sprintf("%s-%s", namePrefix, "vm-custom-ip") | Eventually(func() error { > return AssignIPToVMIP(f, ns, vmipName) | }).WithTimeout(LongWaitDuration).WithPolling(Interval).Should(Succeed()) | }) reflect.Value.call({0x101cd2f00?, 0x140007d15c0?, 0x14000100808?}, {0x101834ea6, 0x4}, {0x1036fb668, 0x0, 0x10011dcac?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/reflect/value.go:581 reflect.Value.Call({0x101cd2f00?, 0x140007d15c0?, 0x140002c0940?}, {0x1036fb668?, 0x101cd2f00?, 0x140007d15c0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/reflect/value.go:365 github.com/onsi/gomega/internal.(*AsyncAssertion).buildActualPoller.func3() /Users/antont/go/pkg/mod/github.com/onsi/gomega@v1.36.1/internal/async_assertion.go:325 github.com/onsi/gomega/internal.(*AsyncAssertion).match(0x140002c0850, {0x1021fecc8, 0x1036fb668}, 0x1, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/gomega@v1.36.1/internal/async_assertion.go:398 github.com/onsi/gomega/internal.(*AsyncAssertion).Should(0x140002c0850, {0x1021fecc8, 0x1036fb668}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/gomega@v1.36.1/internal/async_assertion.go:145 > github.com/deckhouse/virtualization/tests/e2e.init.func11.7.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 | Eventually(func() error { | return AssignIPToVMIP(f, ns, vmipName) > }).WithTimeout(LongWaitDuration).WithPolling(Interval).Should(Succeed()) | }) | github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 8 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 116 [sync.Cond.Wait] sync.runtime_notifyListWait(0x1400087edc8, 0x2ff) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x1400087edb8) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x1400087edb0, {0x140009d7000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x140009d0a20?}, {0x140009d7000?, 0x101cc9640?, 0x140008b0e70?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x14000819440) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x14000819440) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x140009a0608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x140009a0608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x140009a0608, {0x14000b6d285, 0x92d7b, 0x100231c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x1400029a008, {0x14000b6d285, 0x92d7b, 0x92d7b}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000b6d285?, 0x140008b7108?, 0x14000b6d158?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x1400099bda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x140009d0480, {0x12fa5dc00, 0x1400090cb40}, {0x1021ed920, 0x14000075860}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 99 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | < Exit [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 @ 10/09/25 09:14:22.738 (1.207s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 09:14:22.738 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/09/25 09:14:27.516 (4.777s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400077f590>: the container "virtualization-controller" was restarted: virtualization-controller-669d7897c9-zsqpn the container "virtualization-controller" was not found: virtualization-controller-98cf87c7-n9w9c { errs: [ <*errors.joinError | 0x1400077f578>{ errs: [ <*errors.errorString | 0x1400062fe50>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-669d7897c9-zsqpn", }, ], }, <*errors.errorString | 0x1400062fe80>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-98cf87c7-n9w9c", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/09/25 09:14:27.775 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 09:14:27.519 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 09:14:27.519 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 09:14:27.519 [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400077f590>: the container "virtualization-controller" was restarted: virtualization-controller-669d7897c9-zsqpn the container "virtualization-controller" was not found: virtualization-controller-98cf87c7-n9w9c { errs: [ <*errors.joinError | 0x1400077f578>{ errs: [ <*errors.errorString | 0x1400062fe50>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-669d7897c9-zsqpn", }, ], }, <*errors.errorString | 0x1400062fe80>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-98cf87c7-n9w9c", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/09/25 09:14:27.775 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/09/25 09:14:27.775 (257ms) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/09/25 09:14:27.776 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/09/25 09:14:40.455 (12.679s) + + + \ No newline at end of file diff --git a/artifacts/test-sds-20251009-221516-17193/junit.xml b/artifacts/test-sds-20251009-221516-17193/junit.xml new file mode 100644 index 0000000000..d8a633a96f --- /dev/null +++ b/artifacts/test-sds-20251009-221516-17193/junit.xml @@ -0,0 +1,2 @@ + + \ No newline at end of file diff --git a/artifacts/test-sds-20251010-084801-24742/junit.xml b/artifacts/test-sds-20251010-084801-24742/junit.xml new file mode 100644 index 0000000000..b4fb60af0b --- /dev/null +++ b/artifacts/test-sds-20251010-084801-24742/junit.xml @@ -0,0 +1,843 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/10/25 09:09:46.446 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/10/25 09:09:51.27 (4.823s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/10/25 09:09:51.27 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/10/25 09:09:51.27 (0s) + + + + > Enter [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/10/25 09:09:51.275 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:41 @ 10/10/25 09:09:51.275 < Exit [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/10/25 09:09:51.275 (1ms) > Enter [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/10/25 09:09:51.275 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/10/25 09:09:51.275 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/10/25 09:09:55.785 (4.509s) < Exit [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/10/25 09:09:55.785 (4.509s) + + + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/10/25 09:09:55.785 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/10/25 09:09:57.713 (1.928s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/10/25 09:09:57.713 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/10/25 09:10:01.984 (4.271s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/10/25 09:10:01.984 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/10/25 09:10:01.985 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/10/25 09:10:01.985 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:97 @ 10/10/25 09:10:01.985 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/10/25 09:10:19.487 (17.501s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/10/25 09:10:19.487 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/10/25 09:10:19.487 (0s) + + + [FAILED] Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-connectivity-b --namespace head-45bdb83d-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-connectivity-b\n", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-connectivity-a --namespace head-45bdb83d-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1010 09:36:04.654467 4542 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-connectivity-a\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 @ 10/10/25 09:39:33.879 + > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/10/25 09:10:19.487 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 @ 10/10/25 09:10:19.487 [FAILED] Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-connectivity-b --namespace head-45bdb83d-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-connectivity-b\n", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-connectivity-a --namespace head-45bdb83d-end-to-end-connectivity --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1010 09:36:04.654467 4542 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-connectivity-a\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 @ 10/10/25 09:39:33.879 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/10/25 09:39:33.879 (16m42.128s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/10/25 09:39:33.879 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/10/25 09:39:38.978 (5.098s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:118 @ 10/10/25 09:39:38.978 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:129 @ 10/10/25 09:39:38.978 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:142 @ 10/10/25 09:39:38.978 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:166 @ 10/10/25 09:39:38.978 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:174 @ 10/10/25 09:39:38.978 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:179 @ 10/10/25 09:39:38.978 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:187 @ 10/10/25 09:39:38.978 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:198 @ 10/10/25 09:39:38.978 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:209 @ 10/10/25 09:39:38.978 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:222 @ 10/10/25 09:39:38.978 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:228 @ 10/10/25 09:39:38.978 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:240 @ 10/10/25 09:39:38.978 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:250 @ 10/10/25 09:39:38.978 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:258 @ 10/10/25 09:39:38.978 + + + + > Enter [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/10/25 09:39:38.978 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:51 @ 10/10/25 09:39:38.978 < Exit [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/10/25 09:39:38.978 (0s) > Enter [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/10/25 09:39:38.978 < Exit [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/10/25 09:39:38.978 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:82 @ 10/10/25 09:39:38.978 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:94 @ 10/10/25 09:39:38.979 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:106 @ 10/10/25 09:39:38.979 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:116 @ 10/10/25 09:39:38.979 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:132 @ 10/10/25 09:39:38.979 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:151 @ 10/10/25 09:39:38.979 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:157 @ 10/10/25 09:39:38.979 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:189 @ 10/10/25 09:39:38.979 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:216 @ 10/10/25 09:39:38.979 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:252 @ 10/10/25 09:39:38.979 + + + [FAILED] Warning: resource namespaces/testcases is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/vm-disk-attachment": admission webhook "vd.virtualization-controller.validate.d8-virtualization" denied the request: the VirtualDisk name "pr-number-or-commit-hash-vd-attach-automatic-with-hotplug-standalone" is too long: it must be no more than 60 characters Error from server (Forbidden): error when creating "/tmp/testdata/vm-disk-attachment": admission webhook "vd.virtualization-controller.validate.d8-virtualization" denied the request: the VirtualDisk name "pr-number-or-commit-hash-vd-root-automatic-with-hotplug-standalone" is too long: it must be no more than 60 characters Error from server (Forbidden): error when creating "/tmp/testdata/vm-disk-attachment": admission webhook "vm.virtualization-controller.validate.d8-virtualization" denied the request: VirtualDisk name "pr-number-or-commit-hash-vd-root-automatic-with-hotplug-standalone" is too long: it must be no more than 60 characters Expected <bool>: false to equal <bool>: true In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:82 @ 10/10/25 09:39:44.177 + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/10/25 09:39:38.979 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/10/25 09:39:41.011 (2.032s) > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/10/25 09:39:41.011 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/10/25 09:39:41.011 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/10/25 09:39:41.011 [FAILED] Warning: resource namespaces/testcases is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/vm-disk-attachment": admission webhook "vd.virtualization-controller.validate.d8-virtualization" denied the request: the VirtualDisk name "pr-number-or-commit-hash-vd-attach-automatic-with-hotplug-standalone" is too long: it must be no more than 60 characters Error from server (Forbidden): error when creating "/tmp/testdata/vm-disk-attachment": admission webhook "vd.virtualization-controller.validate.d8-virtualization" denied the request: the VirtualDisk name "pr-number-or-commit-hash-vd-root-automatic-with-hotplug-standalone" is too long: it must be no more than 60 characters Error from server (Forbidden): error when creating "/tmp/testdata/vm-disk-attachment": admission webhook "vm.virtualization-controller.validate.d8-virtualization" denied the request: VirtualDisk name "pr-number-or-commit-hash-vd-root-automatic-with-hotplug-standalone" is too long: it must be no more than 60 characters Expected <bool>: false to equal <bool>: true In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:82 @ 10/10/25 09:39:44.177 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/10/25 09:39:44.177 (3.166s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/10/25 09:39:44.177 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/10/25 09:39:48.868 (4.691s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/10/25 09:39:48.868 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/10/25 09:39:48.868 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/10/25 09:39:48.868 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/10/25 09:39:48.868 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/10/25 09:39:48.868 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/10/25 09:39:48.868 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/10/25 09:39:48.868 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/10/25 09:39:48.868 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/10/25 09:39:48.868 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/10/25 09:39:48.868 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/10/25 09:39:48.868 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/10/25 09:39:48.868 + + + > Enter [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/10/25 09:39:48.868 < Exit [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/10/25 09:39:50.794 (1.926s) > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/10/25 09:39:50.794 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/10/25 09:39:50.794 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/10/25 09:39:50.794 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/10/25 09:39:57.506 (6.712s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/10/25 09:39:57.506 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/10/25 09:39:57.506 (0s) + + + [FAILED] Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io vm-restore-force-vm-restore-force --namespace test-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/vm-restore-force-vm-restore-force\n", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io vm-restore-force-vm-always-on-restore-force --namespace test-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/vm-restore-force-vm-always-on-restore-force\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 @ 10/10/25 09:56:39.682 + > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/10/25 09:39:57.507 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/10/25 09:39:57.507 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/10/25 09:39:57.507 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/10/25 09:39:57.507 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/10/25 09:56:39.682 (16m42.168s) [FAILED] Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io vm-restore-force-vm-restore-force --namespace test-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/vm-restore-force-vm-restore-force\n", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io vm-restore-force-vm-always-on-restore-force --namespace test-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/vm-restore-force-vm-always-on-restore-force\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 @ 10/10/25 09:56:39.682 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/10/25 09:56:39.682 (16m42.169s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/10/25 09:56:39.682 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/10/25 09:56:44.607 (4.925s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:131 @ 10/10/25 09:56:44.607 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:177 @ 10/10/25 09:56:44.607 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:341 @ 10/10/25 09:56:44.607 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:367 @ 10/10/25 09:56:44.607 + + + [FAILED] kubectl create namespace testcases Unexpected error: <*exec.ExitError | 0x140003bb840>: exit status 1 { ProcessState: { pid: 6150, status: 256, rusage: { Utime: { Sec: 0, Usec: 29061, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 12679, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47333376, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3692, Majflt: 9, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 13, Nsignals: 33, Nvcsw: 14, Nivcsw: 737, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:48 @ 10/10/25 09:56:45.126 + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/10/25 09:56:44.607 [FAILED] kubectl create namespace testcases Unexpected error: <*exec.ExitError | 0x140003bb840>: exit status 1 { ProcessState: { pid: 6150, status: 256, rusage: { Utime: { Sec: 0, Usec: 29061, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 12679, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47333376, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3692, Majflt: 9, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 13, Nsignals: 33, Nvcsw: 14, Nivcsw: 737, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:48 @ 10/10/25 09:56:45.126 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/10/25 09:56:45.126 (519ms) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/10/25 09:56:45.126 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/10/25 09:56:50.569 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/10/25 09:56:55.049 (4.48s) < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/10/25 09:56:55.049 (9.923s) + + + [FAILED] kubectl create namespace testcases Unexpected error: <*exec.ExitError | 0x140001b49a0>: exit status 1 { ProcessState: { pid: 6194, status: 256, rusage: { Utime: { Sec: 0, Usec: 26926, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 10717, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47153152, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3679, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 49, Nvcsw: 38, Nivcsw: 756, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:64 @ 10/10/25 09:56:55.792 + > Enter [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/10/25 09:56:55.049 [FAILED] kubectl create namespace testcases Unexpected error: <*exec.ExitError | 0x140001b49a0>: exit status 1 { ProcessState: { pid: 6194, status: 256, rusage: { Utime: { Sec: 0, Usec: 26926, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 10717, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 47153152, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 3679, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 13, Msgrcv: 12, Nsignals: 49, Nvcsw: 38, Nivcsw: 756, }, }, Stderr: nil, } occurred In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:64 @ 10/10/25 09:56:55.792 < Exit [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/10/25 09:56:55.792 (743ms) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/10/25 09:56:55.792 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/10/25 09:57:00.388 (4.595s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/10/25 09:57:00.388 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/10/25 09:57:00.388 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/10/25 09:57:00.388 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:126 @ 10/10/25 09:57:00.388 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:140 @ 10/10/25 09:57:00.388 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:152 @ 10/10/25 09:57:00.388 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:157 @ 10/10/25 09:57:00.388 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:163 @ 10/10/25 09:57:00.388 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:176 @ 10/10/25 09:57:00.388 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:181 @ 10/10/25 09:57:00.388 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:187 @ 10/10/25 09:57:00.388 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:202 @ 10/10/25 09:57:00.388 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:216 @ 10/10/25 09:57:00.388 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:240 @ 10/10/25 09:57:00.389 + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:66 @ 10/10/25 09:57:02.384 + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:54 @ 10/10/25 09:57:00.389 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:66 @ 10/10/25 09:57:02.384 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:54 @ 10/10/25 09:57:02.385 (1.996s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/10/25 09:57:02.385 < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/10/25 09:57:06.945 (4.56s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:95 @ 10/10/25 09:57:06.945 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:106 @ 10/10/25 09:57:06.946 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:117 @ 10/10/25 09:57:06.946 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:128 @ 10/10/25 09:57:06.946 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:139 @ 10/10/25 09:57:06.946 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:159 @ 10/10/25 09:57:06.946 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:180 @ 10/10/25 09:57:06.946 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:224 @ 10/10/25 09:57:06.946 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:285 @ 10/10/25 09:57:06.946 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:331 @ 10/10/25 09:57:06.946 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:357 @ 10/10/25 09:57:06.946 + + + + > Enter [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/10/25 09:57:06.947 [SKIPPED] Module SDN is disabled. Skipping all tests for module SDN. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:50 @ 10/10/25 09:57:07.492 < Exit [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/10/25 09:57:07.492 (545ms) > Enter [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/10/25 09:57:07.492 < Exit [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/10/25 09:57:07.492 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/10/25 09:57:07.492 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/10/25 09:57:07.492 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:111 @ 10/10/25 09:57:07.492 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:125 @ 10/10/25 09:57:07.493 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:141 @ 10/10/25 09:57:07.493 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:156 @ 10/10/25 09:57:07.493 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:169 @ 10/10/25 09:57:07.493 + + + > Enter [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/10/25 09:57:07.493 < Exit [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/10/25 09:57:09.44 (1.947s) > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/10/25 09:57:09.44 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/10/25 09:57:09.44 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/10/25 09:57:09.44 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/10/25 09:57:13.801 (4.361s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/10/25 09:57:13.801 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/10/25 09:57:13.801 (0s) + + + [FAILED] Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-45bdb83d-vm-restore-safe --namespace head-45bdb83d-end-to-end-vm-restore-safe --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-45bdb83d-vm-restore-safe\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:107 @ 10/10/25 10:13:55.955 + > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/10/25 09:57:13.801 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/10/25 09:57:13.801 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/10/25 09:57:13.801 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/10/25 09:57:13.801 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/10/25 10:13:55.955 (16m42.15s) [FAILED] Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-45bdb83d-vm-restore-safe --namespace head-45bdb83d-end-to-end-vm-restore-safe --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-45bdb83d-vm-restore-safe\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:107 @ 10/10/25 10:13:55.955 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/10/25 10:13:55.955 (16m42.151s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/10/25 10:13:55.955 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/10/25 10:14:00.885 (4.93s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:115 @ 10/10/25 10:14:00.885 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:158 @ 10/10/25 10:14:00.886 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:348 @ 10/10/25 10:14:00.886 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:374 @ 10/10/25 10:14:00.886 + + + > Enter [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/10/25 10:14:00.886 < Exit [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/10/25 10:14:04.312 (3.426s) > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/10/25 10:14:04.312 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/10/25 10:14:04.312 (0s) > Enter [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/10/25 10:14:04.312 STEP: Create a vmip automatically and check its binding with a lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:81 @ 10/10/25 10:14:04.312 STEP: Remove label from the lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:87 @ 10/10/25 10:14:05.033 STEP: Wait for the label to be restored by the controller - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:96 @ 10/10/25 10:14:05.153 < Exit [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/10/25 10:14:05.386 (1.073s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/10/25 10:14:05.386 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/10/25 10:14:05.386 (0s) + + + > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/10/25 10:14:05.386 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/10/25 10:14:05.386 (0s) > Enter [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/10/25 10:14:05.386 STEP: Create an intermediate vmip automatically to allocate a new ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:109 @ 10/10/25 10:14:05.386 STEP: Delete the intermediate vmip automatically and check that the lease is released - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:115 @ 10/10/25 10:14:05.759 STEP: Reuse the released lease with a static vmip - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:124 @ 10/10/25 10:14:05.996 STEP: Delete the static vmip and lease, then create another static vmip with this ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:132 @ 10/10/25 10:14:06.362 < Exit [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/10/25 10:14:06.956 (1.57s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/10/25 10:14:06.956 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/10/25 10:14:06.956 (0s) + + + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:51 @ 10/10/25 10:14:06.957 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:51 @ 10/10/25 10:14:08.949 (1.992s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:61 @ 10/10/25 10:14:08.949 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:61 @ 10/10/25 10:14:36.259 (27.309s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/10/25 10:14:36.259 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/10/25 10:14:36.259 (0s) + + + > Enter [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:82 @ 10/10/25 10:14:36.259 < Exit [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:82 @ 10/10/25 10:14:37.534 (1.275s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/10/25 10:14:37.534 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/10/25 10:14:37.534 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:97 @ 10/10/25 10:14:37.534 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:98 @ 10/10/25 10:14:37.534 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:97 @ 10/10/25 10:14:38.779 (1.245s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/10/25 10:14:38.779 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/10/25 10:14:38.78 (0s) + + + > Enter [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:108 @ 10/10/25 10:14:38.78 STEP: CVIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:109 @ 10/10/25 10:14:38.78 < Exit [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:108 @ 10/10/25 10:14:40.051 (1.271s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/10/25 10:14:40.051 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/10/25 10:14:40.051 (0s) + + + > Enter [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:119 @ 10/10/25 10:14:40.051 STEP: VMClasses should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:120 @ 10/10/25 10:14:40.051 < Exit [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:119 @ 10/10/25 10:14:41.288 (1.237s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/10/25 10:14:41.288 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/10/25 10:14:41.288 (0s) + + + [FAILED] Timed out after 344.536s. Expected success, but got an error: <*errors.errorString | 0x14000648050>: cannot patch VMIP "head-45bdb83d-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-45bdb83d-vm-custom-ip { s: "cannot patch VMIP \"head-45bdb83d-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-45bdb83d-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/10/25 10:20:25.828 + > Enter [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 @ 10/10/25 10:14:41.288 [FAILED] Timed out after 344.536s. Expected success, but got an error: <*errors.errorString | 0x14000648050>: cannot patch VMIP "head-45bdb83d-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-45bdb83d-vm-custom-ip { s: "cannot patch VMIP \"head-45bdb83d-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-45bdb83d-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/10/25 10:20:25.828 < Exit [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 @ 10/10/25 10:20:25.83 (5m44.538s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/10/25 10:20:25.83 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/10/25 10:20:30.902 (5.072s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:137 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:148 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:158 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:169 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:180 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:192 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:216 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:240 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:263 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:275 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:293 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:310 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:334 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:347 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:351 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:368 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:381 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:398 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:437 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:472 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:490 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:505 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:521 @ 10/10/25 10:20:30.902 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:538 @ 10/10/25 10:20:30.902 + + + [FAILED] Warning: resource namespaces/head-45bdb83d-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "nfs" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x140001b48c0>: exit status 1 { ProcessState: { pid: 7105, status: 256, rusage: { Utime: { Sec: 0, Usec: 277661, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 59542, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 69730304, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5067, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 49, Msgrcv: 56, Nsignals: 356, Nvcsw: 281, Nivcsw: 2624, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/10/25 10:20:38.186 + > Enter [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/10/25 10:20:30.902 < Exit [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/10/25 10:20:34.051 (3.149s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/10/25 10:20:34.051 [FAILED] Warning: resource namespaces/head-45bdb83d-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "nfs" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x140001b48c0>: exit status 1 { ProcessState: { pid: 7105, status: 256, rusage: { Utime: { Sec: 0, Usec: 277661, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 59542, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 69730304, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5067, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 49, Msgrcv: 56, Nsignals: 356, Nvcsw: 281, Nivcsw: 2624, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/10/25 10:20:38.186 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/10/25 10:20:38.186 (4.135s) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/10/25 10:20:38.186 < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/10/25 10:20:42.84 (4.654s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:112 @ 10/10/25 10:20:42.841 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:147 @ 10/10/25 10:20:42.841 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:190 @ 10/10/25 10:20:42.841 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:196 @ 10/10/25 10:20:42.841 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:204 @ 10/10/25 10:20:42.841 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:224 @ 10/10/25 10:20:42.841 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:236 @ 10/10/25 10:20:42.841 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:258 @ 10/10/25 10:20:42.841 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:284 @ 10/10/25 10:20:42.841 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:294 @ 10/10/25 10:20:42.841 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:309 @ 10/10/25 10:20:42.841 + + + > Enter [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/10/25 10:20:42.841 < Exit [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/10/25 10:20:44.771 (1.929s) > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/10/25 10:20:44.771 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/10/25 10:20:44.771 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/10/25 10:20:44.771 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/10/25 10:20:46.756 (1.985s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/10/25 10:20:46.756 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/10/25 10:20:46.756 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/10/25 10:20:46.756 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/10/25 10:20:46.756 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/10/25 10:20:46.756 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:68 @ 10/10/25 10:20:46.756 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/10/25 10:21:02.38 (15.624s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/10/25 10:21:02.381 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/10/25 10:21:02.381 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/10/25 10:21:02.381 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/10/25 10:21:02.381 (0s) > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/10/25 10:21:02.381 STEP: VM should be in Running phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:79 @ 10/10/25 10:21:02.381 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/10/25 10:21:05.421 (3.04s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/10/25 10:21:05.421 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/10/25 10:21:05.421 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/10/25 10:21:05.421 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/10/25 10:21:05.421 (0s) > Enter [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/10/25 10:21:05.421 < Exit [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/10/25 10:21:05.421 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/10/25 10:21:05.421 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/10/25 10:21:05.421 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/10/25 10:21:05.422 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/10/25 10:21:05.422 (0s) > Enter [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/10/25 10:21:05.422 < Exit [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/10/25 10:21:05.422 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/10/25 10:21:05.422 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/10/25 10:21:05.422 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/10/25 10:21:05.422 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/10/25 10:21:05.422 (0s) > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/10/25 10:21:05.422 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/10/25 10:21:05.422 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/10/25 10:22:15.641 (1m10.219s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/10/25 10:22:15.641 (1m10.219s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/10/25 10:22:15.641 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/10/25 10:22:15.641 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.642 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/10/25 10:22:15.642 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.642 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.642 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/10/25 10:22:15.642 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.642 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.642 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/10/25 10:22:15.643 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.643 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.643 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/10/25 10:22:15.643 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.643 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.643 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/10/25 10:22:15.643 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.643 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.643 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/10/25 10:22:15.643 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.643 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.643 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/10/25 10:22:15.643 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.643 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.643 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/10/25 10:22:15.644 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.644 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.644 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/10/25 10:22:15.644 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/10/25 10:22:15.644 (0s) + + + > Enter [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/10/25 10:22:15.644 < Exit [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/10/25 10:22:17.662 (2.018s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/10/25 10:22:17.662 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/10/25 10:22:22.331 (4.668s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/10/25 10:22:22.331 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/10/25 10:22:22.331 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/10/25 10:22:22.332 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:89 @ 10/10/25 10:22:22.332 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/10/25 10:22:28.602 (6.27s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/10/25 10:22:28.602 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/10/25 10:22:28.602 (0s) + + + [FAILED] Expected <[]string | len:2, cap:4>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-root-automatic-conf --namespace head-45bdb83d-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-root-automatic-conf\n", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-root-manual-conf --namespace head-45bdb83d-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-root-manual-conf\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:100 @ 10/10/25 10:39:10.659 + > Enter [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/10/25 10:22:28.602 [FAILED] Expected <[]string | len:2, cap:4>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-root-automatic-conf --namespace head-45bdb83d-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-root-automatic-conf\n", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-root-manual-conf --namespace head-45bdb83d-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-root-manual-conf\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:100 @ 10/10/25 10:39:10.659 < Exit [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/10/25 10:39:10.66 (16m42.168s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/10/25 10:39:10.66 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/10/25 10:39:15.34 (4.68s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:109 @ 10/10/25 10:39:15.34 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:123 @ 10/10/25 10:39:15.34 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:147 @ 10/10/25 10:39:15.34 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:161 @ 10/10/25 10:39:15.34 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:183 @ 10/10/25 10:39:15.34 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:202 @ 10/10/25 10:39:15.34 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:226 @ 10/10/25 10:39:15.34 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:240 @ 10/10/25 10:39:15.34 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:250 @ 10/10/25 10:39:15.34 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:265 @ 10/10/25 10:39:15.34 + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.34 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/10/25 10:39:15.34 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.34 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.34 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/10/25 10:39:15.34 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.34 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.34 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/10/25 10:39:15.34 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.34 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.34 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/10/25 10:39:15.34 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.34 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.341 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/10/25 10:39:15.341 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.341 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.341 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/10/25 10:39:15.341 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.341 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.341 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/10/25 10:39:15.341 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.341 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.341 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/10/25 10:39:15.341 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.341 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.341 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/10/25 10:39:15.341 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.341 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.341 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/10/25 10:39:15.341 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.341 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.341 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/10/25 10:39:15.341 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.341 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.341 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/10/25 10:39:15.341 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/10/25 10:39:15.341 (0s) + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:50 @ 10/10/25 10:39:16.825 + > Enter [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:38 @ 10/10/25 10:39:15.341 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:50 @ 10/10/25 10:39:16.825 < Exit [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:38 @ 10/10/25 10:39:16.826 (1.484s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/10/25 10:39:16.826 < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/10/25 10:39:21.243 (4.418s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:87 @ 10/10/25 10:39:21.244 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:96 @ 10/10/25 10:39:21.244 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:107 @ 10/10/25 10:39:21.244 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:116 @ 10/10/25 10:39:21.244 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:127 @ 10/10/25 10:39:21.244 + + + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/10/25 10:39:21.244 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/10/25 10:39:23.211 (1.967s) > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/10/25 10:39:23.211 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/10/25 10:39:23.211 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/10/25 10:39:23.211 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/10/25 10:39:26.231 (3.021s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/10/25 10:39:26.231 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/10/25 10:39:26.231 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/10/25 10:39:26.232 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/10/25 10:39:26.232 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/10/25 10:39:26.232 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:75 @ 10/10/25 10:39:26.232 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/10/25 10:39:31.231 (4.999s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/10/25 10:39:31.231 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/10/25 10:39:31.231 (0s) + + + [FAILED] Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-root-label-annotation --namespace head-45bdb83d-end-to-end-vm-label-annotation --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-root-label-annotation\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 @ 10/10/25 10:56:13.325 + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/10/25 10:39:31.231 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/10/25 10:39:31.231 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/10/25 10:39:31.231 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 @ 10/10/25 10:39:31.231 [FAILED] Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-root-label-annotation --namespace head-45bdb83d-end-to-end-vm-label-annotation --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-root-label-annotation\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 @ 10/10/25 10:56:13.325 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/10/25 10:56:13.326 (16m42.104s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/10/25 10:56:13.326 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/10/25 10:56:17.924 (4.599s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/10/25 10:56:17.924 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/10/25 10:56:17.924 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/10/25 10:56:17.924 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/10/25 10:56:17.924 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/10/25 10:56:17.924 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/10/25 10:56:17.925 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/10/25 10:56:17.925 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/10/25 10:56:17.925 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/10/25 10:56:17.925 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/10/25 10:56:17.925 + + + > Enter [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/10/25 10:56:17.925 < Exit [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/10/25 10:56:17.925 (0s) > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/10/25 10:56:17.925 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/10/25 10:56:17.925 (0s) > Enter [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/10/25 10:56:17.925 < Exit [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/10/25 10:56:19.405 (1.48s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/10/25 10:56:19.405 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/10/25 10:56:19.406 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/10/25 10:56:19.406 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/10/25 10:56:19.406 (0s) > Enter [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/10/25 10:56:19.406 STEP: Project should be deployed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:73 @ 10/10/25 10:56:19.406 < Exit [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/10/25 10:56:22.127 (2.721s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/10/25 10:56:22.127 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/10/25 10:56:22.127 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/10/25 10:56:22.127 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/10/25 10:56:22.127 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/10/25 10:56:22.127 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/10/25 10:56:25.061 (2.934s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/10/25 10:56:25.061 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/10/25 10:56:25.061 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/10/25 10:56:25.061 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/10/25 10:56:25.061 (0s) > Enter [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/10/25 10:56:25.061 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/10/25 10:56:25.061 < Exit [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/10/25 10:56:29.815 (4.754s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/10/25 10:56:29.815 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/10/25 10:56:29.815 (0s) + + + [FAILED] Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-cvi --namespace head-45bdb83d-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-cvi\n", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd --namespace head-45bdb83d-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/10/25 11:13:11.955 + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/10/25 10:56:29.816 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/10/25 10:56:29.816 (0s) > Enter [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/10/25 10:56:29.816 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/10/25 10:56:29.816 [FAILED] Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd-cvi --namespace head-45bdb83d-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd-cvi\n", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-45bdb83d-vd --namespace head-45bdb83d-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-45bdb83d-vd\n", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/10/25 11:13:11.955 < Exit [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/10/25 11:13:11.955 (16m42.114s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/10/25 11:13:11.955 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/10/25 11:13:17.136 (5.181s) > Enter [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/10/25 11:13:17.136 STEP: Delete manifests - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:50 @ 10/10/25 11:13:17.136 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/10/25 11:13:17.136 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/10/25 11:13:19.823 (2.687s) < Exit [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/10/25 11:13:19.823 (2.687s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/10/25 11:13:19.823 + + + > Enter [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/10/25 11:13:19.823 < Exit [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/10/25 11:13:21.789 (1.966s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/10/25 11:13:21.789 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/10/25 11:13:27.426 (5.636s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/10/25 11:13:27.426 < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/10/25 11:13:27.426 (0s) + + + [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/10/25 11:22:11.469 This is the Progress Report generated when the suite timeout occurred: VirtualMachineMigration When virtual machines are applied checks VMs phases (Spec Runtime: 8m44.045s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 In [It] (Node Runtime: 8m44.045s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 At [By Step] Virtual machine agents should be ready (Step Runtime: 8m44.045s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:76 Spec Goroutine goroutine 701 [sync.WaitGroup.Wait, 10 minutes] sync.runtime_SemacquireWaitGroup(0x104cf4b30?, 0x40?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x140008eeab0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140001b4ea0, 0x2, 0x106764e70?}, {0x1064830d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400067c390, {0x14000954540, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1064830d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400067c390, {0x14000954540, 0x25}, {0x14000c54180, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1064830d7, 0x2a}, {0x10644037b, 0x7}, {{0x0, 0x0, 0x0}, 0x1400067c390, {0x14000954540, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x1400067c390, {0x14000954540, 0x25}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func20.4.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:77 | It("checks VMs phases", func() { | By("Virtual machine agents should be ready") > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x1063e4050?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 6 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 750 [syscall, 10 minutes] syscall.syscall6(0x1008eeaf0?, 0x1149cb0c8?, 0x10904cf30?, 0x90?, 0x14000301808?, 0x14000ac63f0?, 0x1400089fa68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400089fa98?, 0x104d766fc?, 0x90?, 0x106d64660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400034aa10?, 0x1400089fad4, 0x140008eeaf0?, 0x1400034a9a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140004907c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000a1b008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x1400093c300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x1400093c300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400031b470?, 0x106dfc1e8?, 0x1400034a850?}}, {0x106e1c808?, 0x1400034a850?}, {0x14000a1e000?, 0xc2324ee91c091fc0?}, {0x106df5120, 0x14000490740}, {0x106df5120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400031b470?, 0x108303668?, 0x1082d9ee0?}}, {0x106e1c808, 0x1400034a850}, {0x14000a1e000, 0xc3}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106e20e48?, 0x1400000fea8?}, {0x106440924?, 0x1400062d6d8?}}, {0x1064830d7, 0x2a}, {0x1400092e5e0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 701 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 749 [syscall, 10 minutes] syscall.syscall6(0x1009ae040?, 0x1149cbca8?, 0x10904d3e8?, 0x90?, 0x140000a5808?, 0x14000916090?, 0x14000b19a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000b19a98?, 0x104d766fc?, 0x90?, 0x106d64660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002e8700?, 0x14000b19ad4, 0x140009ae040?, 0x140002e85b0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140008040c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000922408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000920000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000920000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400031b470?, 0x106dfc1e8?, 0x140002e8000?}}, {0x106e1c808?, 0x140002e8000?}, {0x1400091e000?, 0xc2324ee91c095670?}, {0x106df5120, 0x14000804000}, {0x106df5120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400031b470?, 0x108303668?, 0x1082d9ee0?}}, {0x106e1c808, 0x140002e8000}, {0x1400091e000, 0xc3}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106e20e48?, 0x1400000fea8?}, {0x106440924?, 0x140009a96d8?}}, {0x1064830d7, 0x2a}, {0x1400092e5c0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 701 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) + > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/10/25 11:13:27.426 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:76 @ 10/10/25 11:13:27.426 [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/10/25 11:22:11.469 This is the Progress Report generated when the suite timeout occurred: VirtualMachineMigration When virtual machines are applied checks VMs phases (Spec Runtime: 8m44.045s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 In [It] (Node Runtime: 8m44.045s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 At [By Step] Virtual machine agents should be ready (Step Runtime: 8m44.045s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:76 Spec Goroutine goroutine 701 [sync.WaitGroup.Wait, 10 minutes] sync.runtime_SemacquireWaitGroup(0x104cf4b30?, 0x40?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x140008eeab0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x140001b4ea0, 0x2, 0x106764e70?}, {0x1064830d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400067c390, {0x14000954540, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1064830d7, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400067c390, {0x14000954540, 0x25}, {0x14000c54180, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1064830d7, 0x2a}, {0x10644037b, 0x7}, {{0x0, 0x0, 0x0}, 0x1400067c390, {0x14000954540, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x1400067c390, {0x14000954540, 0x25}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func20.4.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:77 | It("checks VMs phases", func() { | By("Virtual machine agents should be ready") > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x1063e4050?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 6 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 750 [syscall, 10 minutes] syscall.syscall6(0x1008eeaf0?, 0x1149cb0c8?, 0x10904cf30?, 0x90?, 0x14000301808?, 0x14000ac63f0?, 0x1400089fa68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400089fa98?, 0x104d766fc?, 0x90?, 0x106d64660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400034aa10?, 0x1400089fad4, 0x140008eeaf0?, 0x1400034a9a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140004907c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000a1b008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x1400093c300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x1400093c300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400031b470?, 0x106dfc1e8?, 0x1400034a850?}}, {0x106e1c808?, 0x1400034a850?}, {0x14000a1e000?, 0xc2324ee91c091fc0?}, {0x106df5120, 0x14000490740}, {0x106df5120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400031b470?, 0x108303668?, 0x1082d9ee0?}}, {0x106e1c808, 0x1400034a850}, {0x14000a1e000, 0xc3}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106e20e48?, 0x1400000fea8?}, {0x106440924?, 0x1400062d6d8?}}, {0x1064830d7, 0x2a}, {0x1400092e5e0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 701 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 749 [syscall, 10 minutes] syscall.syscall6(0x1009ae040?, 0x1149cbca8?, 0x10904d3e8?, 0x90?, 0x140000a5808?, 0x14000916090?, 0x14000b19a68?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000b19a98?, 0x104d766fc?, 0x90?, 0x106d64660?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002e8700?, 0x14000b19ad4, 0x140009ae040?, 0x140002e85b0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140008040c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000922408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000920000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000920000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400031b470?, 0x106dfc1e8?, 0x140002e8000?}}, {0x106e1c808?, 0x140002e8000?}, {0x1400091e000?, 0xc2324ee91c095670?}, {0x106df5120, 0x14000804000}, {0x106df5120, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400031b470?, 0x108303668?, 0x1082d9ee0?}}, {0x106e1c808, 0x140002e8000}, {0x1400091e000, 0xc3}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x106e20e48?, 0x1400000fea8?}, {0x106440924?, 0x140009a96d8?}}, {0x1064830d7, 0x2a}, {0x1400092e5c0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:259 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 701 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/10/25 11:22:11.473 (8m44.048s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/10/25 11:22:11.473 < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/10/25 11:22:16.698 (5.225s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400000f290>: the container "virtualization-controller" was restarted: virtualization-controller-5f7744444b-2lsv9 the container "virtualization-controller" was restarted: virtualization-controller-5f7744444b-crgqz the container "virtualization-controller" was not found: virtualization-controller-5fb7f4bc7b-sfqb2 { errs: [ <*errors.joinError | 0x1400000f278>{ errs: [ <*errors.joinError | 0x1400000f260>{ errs: [ <*errors.errorString | 0x140006678a0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-5f7744444b-2lsv9", }, ], }, <*errors.errorString | 0x140006678e0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-5f7744444b-crgqz", }, ], }, <*errors.errorString | 0x14000667900>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-5fb7f4bc7b-sfqb2", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/10/25 11:22:17.288 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/10/25 11:22:16.698 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/10/25 11:22:16.698 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/10/25 11:22:16.698 [FAILED] Expected success, but got an error: <*errors.joinError | 0x1400000f290>: the container "virtualization-controller" was restarted: virtualization-controller-5f7744444b-2lsv9 the container "virtualization-controller" was restarted: virtualization-controller-5f7744444b-crgqz the container "virtualization-controller" was not found: virtualization-controller-5fb7f4bc7b-sfqb2 { errs: [ <*errors.joinError | 0x1400000f278>{ errs: [ <*errors.joinError | 0x1400000f260>{ errs: [ <*errors.errorString | 0x140006678a0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-5f7744444b-2lsv9", }, ], }, <*errors.errorString | 0x140006678e0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-5f7744444b-crgqz", }, ], }, <*errors.errorString | 0x14000667900>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-5fb7f4bc7b-sfqb2", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/10/25 11:22:17.288 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/10/25 11:22:17.288 (589ms) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/10/25 11:22:17.288 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/10/25 11:22:29.72 (12.432s) + + + \ No newline at end of file diff --git a/artifacts/test-sds-20251011-084506-27483/junit.xml b/artifacts/test-sds-20251011-084506-27483/junit.xml new file mode 100644 index 0000000000..0b915b6fd2 --- /dev/null +++ b/artifacts/test-sds-20251011-084506-27483/junit.xml @@ -0,0 +1,838 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/11/25 10:00:11.688 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/11/25 10:00:16.482 (4.794s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/11/25 10:00:16.482 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/11/25 10:00:16.482 (0s) + + + > Enter [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/11/25 10:00:16.483 < Exit [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/11/25 10:00:17.935 (1.452s) > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/11/25 10:00:17.935 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/11/25 10:00:17.935 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/11/25 10:00:17.935 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/11/25 10:00:24.302 (6.368s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/11/25 10:00:24.302 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/11/25 10:00:24.303 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-always-on-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-always-on-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 @ 10/11/25 10:55:16.602 + > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/11/25 10:00:24.303 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/11/25 10:00:24.303 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/11/25 10:00:24.303 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/11/25 10:00:24.303 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/11/25 10:55:16.602 (16m42.153s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-always-on-restore-force --namespace head-5073ae15-end-to-end-vm-restore-force --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-always-on-restore-force\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 @ 10/11/25 10:55:16.602 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/11/25 10:55:16.602 (16m42.153s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/11/25 10:55:16.602 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/11/25 10:55:27.021 (10.419s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:131 @ 10/11/25 10:55:27.021 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:177 @ 10/11/25 10:55:27.022 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:341 @ 10/11/25 10:55:27.022 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:367 @ 10/11/25 10:55:27.022 + + + > Enter [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/11/25 10:55:27.022 < Exit [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/11/25 10:55:28.964 (1.942s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/11/25 10:55:28.964 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/11/25 10:55:34.305 (5.341s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/11/25 10:55:34.305 < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/11/25 10:55:34.305 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:77 @ 10/11/25 11:12:22.183 + > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/11/25 10:55:34.306 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:76 @ 10/11/25 10:55:34.306 [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-migration --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:77 @ 10/11/25 11:12:22.183 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/11/25 11:12:22.183 (16m42.107s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/11/25 11:12:22.183 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/11/25 11:12:28.299 (6.116s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:86 @ 10/11/25 11:12:28.3 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:100 @ 10/11/25 11:12:28.3 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:116 @ 10/11/25 11:12:28.3 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:131 @ 10/11/25 11:12:28.3 + + + [FAILED] Warning: resource namespaces/head-5073ae15-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "ceph-pool-r2-csi-rbd" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x14000413720>: exit status 1 { ProcessState: { pid: 40942, status: 256, rusage: { Utime: { Sec: 0, Usec: 296394, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 61884, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 70926336, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5146, Majflt: 4, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 49, Msgrcv: 57, Nsignals: 342, Nvcsw: 285, Nivcsw: 2755, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/11/25 11:12:34.994 + > Enter [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/11/25 11:12:28.3 < Exit [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/11/25 11:12:31.299 (2.999s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/11/25 11:12:31.299 [FAILED] Warning: resource namespaces/head-5073ae15-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "ceph-pool-r2-csi-rbd" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x14000413720>: exit status 1 { ProcessState: { pid: 40942, status: 256, rusage: { Utime: { Sec: 0, Usec: 296394, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 61884, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 70926336, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5146, Majflt: 4, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 49, Msgrcv: 57, Nsignals: 342, Nvcsw: 285, Nivcsw: 2755, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/11/25 11:12:34.994 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/11/25 11:12:34.994 (3.695s) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/11/25 11:12:34.994 The list of pods is empty; nothing to dump. < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/11/25 11:12:40.986 (5.992s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:112 @ 10/11/25 11:12:40.986 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:147 @ 10/11/25 11:12:40.986 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:190 @ 10/11/25 11:12:40.987 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:196 @ 10/11/25 11:12:40.987 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:204 @ 10/11/25 11:12:40.987 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:224 @ 10/11/25 11:12:40.987 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:236 @ 10/11/25 11:12:40.987 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:258 @ 10/11/25 11:12:40.987 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:284 @ 10/11/25 11:12:40.987 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:294 @ 10/11/25 11:12:40.987 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:309 @ 10/11/25 11:12:40.987 + + + > Enter [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/11/25 11:12:40.987 < Exit [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/11/25 11:12:40.989 (2ms) > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 11:12:40.989 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 11:12:40.989 (0s) > Enter [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/11/25 11:12:40.989 < Exit [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/11/25 11:12:42.575 (1.586s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 11:12:42.575 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 11:12:42.575 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 11:12:42.576 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 11:12:42.576 (0s) > Enter [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/11/25 11:12:42.576 STEP: Project should be deployed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:73 @ 10/11/25 11:12:42.576 < Exit [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/11/25 11:12:45.321 (2.745s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 11:12:45.321 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 11:12:45.321 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 11:12:45.322 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 11:12:45.322 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/11/25 11:12:45.322 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/11/25 11:12:48.133 (2.812s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 11:12:48.133 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 11:12:48.133 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 11:12:48.134 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 11:12:48.134 (0s) > Enter [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/11/25 11:12:48.134 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/11/25 11:12:48.134 < Exit [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/11/25 11:12:54.926 (6.792s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 11:12:54.926 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 11:12:54.926 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1011 11:39:49.978593 40956 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualdisks/head-5073ae15-vd\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-cvi --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1011 11:43:13.042626 40955 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-cvi\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/11/25 11:48:30.1 + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 11:12:54.926 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/11/25 11:12:54.926 (0s) > Enter [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/11/25 11:12:54.926 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/11/25 11:12:54.926 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1011 11:39:49.978593 40956 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualdisks/head-5073ae15-vd\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-cvi --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1011 11:43:13.042626 40955 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-cvi\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/11/25 11:48:30.1 < Exit [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/11/25 11:48:30.101 (16m42.135s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 11:48:30.102 The list of pods is empty; nothing to dump. < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/11/25 11:48:36.244 (6.142s) > Enter [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/11/25 11:48:36.244 STEP: Delete manifests - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:50 @ 10/11/25 11:48:36.244 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/11/25 11:48:36.244 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/11/25 11:48:38.376 (2.132s) < Exit [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/11/25 11:48:38.376 (2.132s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/11/25 11:48:38.377 + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 @ 10/11/25 12:05:26.894 + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/11/25 11:48:38.377 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/11/25 11:48:39.861 (1.484s) > Enter [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/11/25 11:48:39.861 < Exit [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/11/25 11:48:44.745 (4.884s) > Enter [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/11/25 11:48:44.745 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 @ 10/11/25 11:48:44.745 [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 @ 10/11/25 12:05:26.894 < Exit [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/11/25 12:05:26.895 (16m42.128s) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/11/25 12:05:26.895 The list of pods is empty; nothing to dump. STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/11/25 12:05:33.123 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/11/25 12:05:37.635 (4.512s) < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/11/25 12:05:37.635 (10.74s) + + + + > Enter [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/11/25 12:05:37.635 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:41 @ 10/11/25 12:05:37.636 < Exit [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/11/25 12:05:37.636 (0s) > Enter [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/11/25 12:05:37.636 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/11/25 12:05:37.636 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/11/25 12:05:42.296 (4.66s) < Exit [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/11/25 12:05:42.296 (4.66s) + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/11/25 12:05:43.81 + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/11/25 12:05:42.296 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/11/25 12:05:43.81 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/11/25 12:05:43.81 (1.513s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/11/25 12:05:43.81 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/11/25 12:05:49.846 (6.036s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:100 @ 10/11/25 12:05:49.846 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:111 @ 10/11/25 12:05:49.846 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:122 @ 10/11/25 12:05:49.846 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:133 @ 10/11/25 12:05:49.846 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:144 @ 10/11/25 12:05:49.846 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:164 @ 10/11/25 12:05:49.846 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:185 @ 10/11/25 12:05:49.846 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:229 @ 10/11/25 12:05:49.846 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:290 @ 10/11/25 12:05:49.847 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:341 @ 10/11/25 12:05:49.847 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:367 @ 10/11/25 12:05:49.847 + + + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/11/25 12:05:49.847 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/11/25 12:05:51.8 (1.953s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:62 @ 10/11/25 12:05:51.8 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:62 @ 10/11/25 12:06:19.271 (27.471s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/11/25 12:06:19.271 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/11/25 12:06:19.272 (0s) + + + > Enter [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/11/25 12:06:19.272 < Exit [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/11/25 12:06:20.723 (1.451s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/11/25 12:06:20.723 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/11/25 12:06:20.723 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/11/25 12:06:20.723 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:102 @ 10/11/25 12:06:20.723 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/11/25 12:06:25.469 (4.746s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/11/25 12:06:25.469 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/11/25 12:06:25.469 (0s) + + + > Enter [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/11/25 12:06:25.469 STEP: CVIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:113 @ 10/11/25 12:06:25.469 < Exit [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/11/25 12:06:28.296 (2.826s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/11/25 12:06:28.296 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/11/25 12:06:28.296 (0s) + + + > Enter [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/11/25 12:06:28.296 STEP: VMClasses should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:124 @ 10/11/25 12:06:28.296 < Exit [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/11/25 12:06:31.013 (2.717s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/11/25 12:06:31.013 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/11/25 12:06:31.013 (0s) + + + [FAILED] Timed out after 344.281s. Expected success, but got an error: <*errors.errorString | 0x14000614070>: cannot patch VMIP "head-5073ae15-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip { s: "cannot patch VMIP \"head-5073ae15-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:138 @ 10/11/25 12:12:15.259 + > Enter [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/11/25 12:06:31.014 [FAILED] Timed out after 344.281s. Expected success, but got an error: <*errors.errorString | 0x14000614070>: cannot patch VMIP "head-5073ae15-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip { s: "cannot patch VMIP \"head-5073ae15-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:138 @ 10/11/25 12:12:15.259 < Exit [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/11/25 12:12:15.259 (5m44.282s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/11/25 12:12:15.259 The list of pods is empty; nothing to dump. < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/11/25 12:12:21.672 (6.413s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:141 @ 10/11/25 12:12:21.673 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:152 @ 10/11/25 12:12:21.673 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:162 @ 10/11/25 12:12:21.673 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:173 @ 10/11/25 12:12:21.673 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:184 @ 10/11/25 12:12:21.673 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:196 @ 10/11/25 12:12:21.673 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:220 @ 10/11/25 12:12:21.673 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:245 @ 10/11/25 12:12:21.673 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:268 @ 10/11/25 12:12:21.673 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:280 @ 10/11/25 12:12:21.674 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:299 @ 10/11/25 12:12:21.674 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:316 @ 10/11/25 12:12:21.674 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:341 @ 10/11/25 12:12:21.674 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:354 @ 10/11/25 12:12:21.674 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:358 @ 10/11/25 12:12:21.674 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:375 @ 10/11/25 12:12:21.674 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:388 @ 10/11/25 12:12:21.674 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:405 @ 10/11/25 12:12:21.674 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:444 @ 10/11/25 12:12:21.674 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:479 @ 10/11/25 12:12:21.674 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:497 @ 10/11/25 12:12:21.674 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:512 @ 10/11/25 12:12:21.675 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:528 @ 10/11/25 12:12:21.675 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:545 @ 10/11/25 12:12:21.675 + + + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/11/25 12:12:21.675 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/11/25 12:12:23.621 (1.946s) > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/11/25 12:12:23.621 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/11/25 12:12:23.621 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/11/25 12:12:23.621 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/11/25 12:12:26.791 (3.17s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/11/25 12:12:26.791 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/11/25 12:12:26.791 (0s) + + + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/11/25 12:12:26.791 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/11/25 12:12:26.791 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/11/25 12:12:26.791 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:88 @ 10/11/25 12:12:26.791 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/11/25 12:12:32.54 (5.749s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/11/25 12:12:32.54 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/11/25 12:12:32.54 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-with-hotplug-standalone --namespace head-5073ae15-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/11/25 12:29:14.623 + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/11/25 12:12:32.541 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/11/25 12:12:32.541 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/11/25 12:12:32.541 STEP: VDs with consumers should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:99 @ 10/11/25 12:12:32.541 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-with-hotplug-standalone --namespace head-5073ae15-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/11/25 12:29:14.623 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/11/25 12:29:14.625 (16m42.111s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/11/25 12:29:14.625 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/11/25 12:29:20.7 (6.075s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/11/25 12:29:20.7 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/11/25 12:29:20.7 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/11/25 12:29:20.7 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/11/25 12:29:20.7 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/11/25 12:29:20.701 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/11/25 12:29:20.701 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/11/25 12:29:20.701 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/11/25 12:29:20.701 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/11/25 12:29:20.701 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/11/25 12:29:20.701 + + + > Enter [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/11/25 12:29:20.701 < Exit [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/11/25 12:29:22.628 (1.927s) > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 12:29:22.628 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 12:29:22.628 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/11/25 12:29:22.628 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/11/25 12:29:24.881 (2.253s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 12:29:24.881 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 12:29:24.881 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 12:29:24.881 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 12:29:24.881 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/11/25 12:29:24.881 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:68 @ 10/11/25 12:29:24.881 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/11/25 12:29:43.385 (18.504s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 12:29:43.385 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 12:29:43.385 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 12:29:43.385 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 12:29:43.385 (0s) > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/11/25 12:29:43.385 STEP: VM should be in Running phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:79 @ 10/11/25 12:29:43.385 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/11/25 12:29:46.424 (3.038s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 12:29:46.424 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 12:29:46.424 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 12:29:46.424 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 12:29:46.424 (0s) > Enter [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/11/25 12:29:46.424 < Exit [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/11/25 12:29:46.424 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 12:29:46.424 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 12:29:46.424 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 12:29:46.424 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 12:29:46.424 (0s) > Enter [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/11/25 12:29:46.424 < Exit [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/11/25 12:29:46.424 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 12:29:46.424 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 12:29:46.424 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 12:29:46.424 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/11/25 12:29:46.424 (0s) > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/11/25 12:29:46.424 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/11/25 12:29:46.424 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/11/25 12:30:56.666 (1m10.242s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/11/25 12:30:56.666 (1m10.242s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 12:30:56.666 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/11/25 12:30:56.666 (0s) + + + + > Enter [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/11/25 12:30:56.667 [SKIPPED] Module SDN is disabled. Skipping all tests for module SDN. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:50 @ 10/11/25 12:30:57.151 < Exit [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/11/25 12:30:57.151 (484ms) > Enter [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/11/25 12:30:57.151 < Exit [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/11/25 12:30:57.151 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/11/25 12:30:57.151 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/11/25 12:30:57.151 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:111 @ 10/11/25 12:30:57.152 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:125 @ 10/11/25 12:30:57.152 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:141 @ 10/11/25 12:30:57.152 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:156 @ 10/11/25 12:30:57.152 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:169 @ 10/11/25 12:30:57.152 + + + + > Enter [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/11/25 12:30:57.152 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:73 @ 10/11/25 12:30:57.152 < Exit [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/11/25 12:30:57.152 (0s) > Enter [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/11/25 12:30:57.152 < Exit [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/11/25 12:30:57.152 (0s) > Enter [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/11/25 12:30:57.152 < Exit [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/11/25 12:30:57.152 (0s) + + + > Enter [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/11/25 12:30:57.153 < Exit [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/11/25 12:30:58.612 (1.46s) > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/11/25 12:30:58.612 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/11/25 12:30:58.612 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/11/25 12:30:58.612 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/11/25 12:31:02.427 (3.814s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/11/25 12:31:02.427 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/11/25 12:31:02.427 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-safe --namespace head-5073ae15-end-to-end-vm-restore-safe --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-safe\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:107 @ 10/11/25 12:47:44.079 + > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/11/25 12:31:02.427 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/11/25 12:31:02.427 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/11/25 12:31:02.427 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/11/25 12:31:02.427 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/11/25 12:47:44.079 (16m41.624s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-safe --namespace head-5073ae15-end-to-end-vm-restore-safe --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-safe\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:107 @ 10/11/25 12:47:44.079 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/11/25 12:47:44.08 (16m41.625s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/11/25 12:47:44.08 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/11/25 12:47:50.238 (6.158s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:115 @ 10/11/25 12:47:50.238 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:158 @ 10/11/25 12:47:50.238 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:348 @ 10/11/25 12:47:50.238 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:374 @ 10/11/25 12:47:50.238 + + + + > Enter [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/11/25 12:47:50.239 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:51 @ 10/11/25 12:47:50.239 < Exit [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/11/25 12:47:50.239 (0s) > Enter [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/11/25 12:47:50.239 < Exit [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/11/25 12:47:50.239 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:82 @ 10/11/25 12:47:50.239 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:94 @ 10/11/25 12:47:50.239 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:106 @ 10/11/25 12:47:50.239 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:116 @ 10/11/25 12:47:50.239 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:132 @ 10/11/25 12:47:50.239 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:151 @ 10/11/25 12:47:50.24 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:157 @ 10/11/25 12:47:50.24 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:189 @ 10/11/25 12:47:50.24 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:216 @ 10/11/25 12:47:50.24 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:252 @ 10/11/25 12:47:50.24 + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.24 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/11/25 12:47:50.24 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.24 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.24 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/11/25 12:47:50.24 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.24 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.24 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/11/25 12:47:50.241 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.241 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.241 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/11/25 12:47:50.241 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.241 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.241 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/11/25 12:47:50.241 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.241 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.241 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/11/25 12:47:50.241 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.241 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.241 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/11/25 12:47:50.241 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.242 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.242 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/11/25 12:47:50.242 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.242 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.242 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/11/25 12:47:50.242 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/11/25 12:47:50.242 (0s) + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:51 @ 10/11/25 12:47:52.219 + > Enter [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/11/25 12:47:50.242 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:51 @ 10/11/25 12:47:52.219 < Exit [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/11/25 12:47:52.219 (1.977s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/11/25 12:47:52.219 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/11/25 12:47:57.816 (5.597s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:94 @ 10/11/25 12:47:57.816 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:103 @ 10/11/25 12:47:57.816 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:114 @ 10/11/25 12:47:57.817 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:128 @ 10/11/25 12:47:57.817 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:139 @ 10/11/25 12:47:57.817 + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.817 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/11/25 12:47:57.817 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.817 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.817 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/11/25 12:47:57.817 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.817 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.817 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/11/25 12:47:57.818 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.818 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.818 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/11/25 12:47:57.818 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.818 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.818 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/11/25 12:47:57.818 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.818 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.818 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/11/25 12:47:57.818 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.818 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.818 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/11/25 12:47:57.818 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.818 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.819 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/11/25 12:47:57.819 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.819 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.819 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/11/25 12:47:57.819 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.819 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.819 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/11/25 12:47:57.819 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.819 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.819 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/11/25 12:47:57.819 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.819 (0s) + + + + > Enter [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.82 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/11/25 12:47:57.82 < Exit [BeforeEach] [sig-storage] LocalVirtualDiskMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/11/25 12:47:57.82 (0s) + + + > Enter [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/11/25 12:47:57.82 < Exit [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/11/25 12:47:59.297 (1.477s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/11/25 12:47:59.297 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/11/25 12:48:03.927 (4.63s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/11/25 12:48:03.927 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/11/25 12:48:03.927 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/11/25 12:48:03.927 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:89 @ 10/11/25 12:48:03.927 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/11/25 12:48:08.907 (4.979s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/11/25 12:48:08.907 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/11/25 12:48:08.907 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:4>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-manual-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-manual-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:100 @ 10/11/25 13:04:51.025 + > Enter [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/11/25 12:48:08.907 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:4>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-manual-conf --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-manual-conf\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:100 @ 10/11/25 13:04:51.025 < Exit [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/11/25 13:04:51.026 (16m42.121s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/11/25 13:04:51.026 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/11/25 13:04:57.251 (6.225s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:109 @ 10/11/25 13:04:57.252 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:123 @ 10/11/25 13:04:57.252 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:147 @ 10/11/25 13:04:57.252 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:161 @ 10/11/25 13:04:57.252 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:183 @ 10/11/25 13:04:57.252 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:202 @ 10/11/25 13:04:57.252 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:226 @ 10/11/25 13:04:57.253 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:240 @ 10/11/25 13:04:57.253 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:250 @ 10/11/25 13:04:57.253 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:265 @ 10/11/25 13:04:57.253 + + + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/11/25 13:04:57.253 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/11/25 13:04:59.195 (1.942s) > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/11/25 13:04:59.195 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/11/25 13:04:59.195 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/11/25 13:04:59.195 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/11/25 13:05:01.788 (2.593s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/11/25 13:05:01.788 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/11/25 13:05:01.788 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/11/25 13:05:01.788 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/11/25 13:05:01.788 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/11/25 13:05:01.788 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:75 @ 10/11/25 13:05:01.788 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/11/25 13:05:06.305 (4.517s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/11/25 13:05:06.305 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/11/25 13:05:06.305 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-label-annotation --namespace head-5073ae15-end-to-end-vm-label-annotation --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-label-annotation\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 @ 10/11/25 14:10:27.595 + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/11/25 13:05:06.305 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/11/25 13:05:06.305 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/11/25 13:05:06.305 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 @ 10/11/25 13:05:06.305 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-label-annotation --namespace head-5073ae15-end-to-end-vm-label-annotation --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-label-annotation\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 @ 10/11/25 14:10:27.595 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/11/25 14:10:27.595 (16m42.158s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/11/25 14:10:27.595 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/11/25 14:10:33.288 (5.693s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/11/25 14:10:33.288 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/11/25 14:10:33.288 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/11/25 14:10:33.289 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/11/25 14:10:33.289 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/11/25 14:10:33.289 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/11/25 14:10:33.289 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/11/25 14:10:33.289 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/11/25 14:10:33.289 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/11/25 14:10:33.289 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/11/25 14:10:33.289 + + + > Enter [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/11/25 14:10:33.289 < Exit [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/11/25 14:10:35.22 (1.931s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/11/25 14:10:35.22 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/11/25 14:10:44.238 (9.017s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/11/25 14:10:44.238 < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/11/25 14:10:44.238 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:6, cap:6>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-affinity --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-affinity\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-b-not-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1011 14:59:22.871391 44466 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-b-not-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-d --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1011 16:29:12.428177 44469 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-d\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-c-and-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1011 16:29:12.427535 44470 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-c-and-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-a-not-b --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-a-not-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-selector --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-selector\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:103 @ 10/11/25 17:02:11.812 + > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/11/25 14:10:44.238 STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/11/25 14:10:44.238 END STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/11/25 14:10:51.099 (6.861s) STEP: `VirtualMachineClasses` should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:95 @ 10/11/25 14:10:51.099 END STEP: `VirtualMachineClasses` should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:95 @ 10/11/25 14:10:53.773 (2.674s) STEP: `VirtualDisks` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:102 @ 10/11/25 14:10:53.773 END STEP: `VirtualDisks` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:102 @ 10/11/25 17:02:11.812 (16m42.118s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:6, cap:6>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-affinity --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-affinity\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-b-not-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1011 14:59:22.871391 44466 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-b-not-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-d --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1011 16:29:12.428177 44469 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-d\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-c-and-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1011 16:29:12.427535 44470 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-c-and-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-a-not-b --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-a-not-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-selector --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-selector\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:103 @ 10/11/25 17:02:11.812 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/11/25 17:02:11.812 (16m51.654s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/11/25 17:02:11.812 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/11/25 17:02:17.373 (5.561s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:120 @ 10/11/25 17:02:17.373 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:283 @ 10/11/25 17:02:17.373 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:372 @ 10/11/25 17:02:17.374 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:467 @ 10/11/25 17:02:17.374 + + + > Enter [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/11/25 17:02:17.374 < Exit [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/11/25 17:02:20.484 (3.11s) > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/11/25 17:02:20.484 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/11/25 17:02:20.484 (0s) > Enter [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/11/25 17:02:20.484 STEP: Create a vmip automatically and check its binding with a lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:81 @ 10/11/25 17:02:20.484 STEP: Remove label from the lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:87 @ 10/11/25 17:02:21.495 STEP: Wait for the label to be restored by the controller - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:96 @ 10/11/25 17:02:21.609 < Exit [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/11/25 17:02:21.837 (1.353s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/11/25 17:02:21.837 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/11/25 17:02:21.837 (0s) + + + > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/11/25 17:02:21.837 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/11/25 17:02:21.838 (0s) > Enter [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/11/25 17:02:21.838 STEP: Create an intermediate vmip automatically to allocate a new ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:109 @ 10/11/25 17:02:21.838 STEP: Delete the intermediate vmip automatically and check that the lease is released - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:115 @ 10/11/25 17:02:22.199 STEP: Reuse the released lease with a static vmip - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:124 @ 10/11/25 17:02:22.425 STEP: Delete the static vmip and lease, then create another static vmip with this ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:132 @ 10/11/25 17:02:22.775 < Exit [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/11/25 17:02:23.355 (1.517s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/11/25 17:02:23.355 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/11/25 17:02:23.355 (0s) + + + > Enter [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/11/25 17:02:23.355 < Exit [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/11/25 17:02:25.353 (1.998s) > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/11/25 17:02:25.353 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/11/25 17:02:25.353 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/11/25 17:02:25.353 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/11/25 17:02:31.674 (6.32s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/11/25 17:02:31.674 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/11/25 17:02:31.674 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/11/25 17:02:31.674 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/11/25 17:02:31.674 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/11/25 17:02:31.674 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:91 @ 10/11/25 17:02:31.674 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/11/25 17:02:36.559 (4.885s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/11/25 17:02:36.559 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/11/25 17:02:36.559 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=WaitForFirstConsumer'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-blank-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=WaitForFirstConsumer' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-blank-not-existing-vmclass-with-changing\n\nwaited for: 'jsonpath={.status.phase}=WaitForFirstConsumer'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=WaitForFirstConsumer' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-not-existing-vmclass-with-changing\n\nwaited for: 'jsonpath={.status.phase}=WaitForFirstConsumer'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:103 @ 10/11/25 17:19:18.666 + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/11/25 17:02:36.56 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/11/25 17:02:36.56 (0s) > Enter [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/11/25 17:02:36.56 STEP: VDs should be in WaitForFirstConsumer phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:102 @ 10/11/25 17:02:36.56 [FAILED] should observe resources in ''jsonpath={.status.phase}=WaitForFirstConsumer'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-blank-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=WaitForFirstConsumer' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-blank-not-existing-vmclass-with-changing\n\nwaited for: 'jsonpath={.status.phase}=WaitForFirstConsumer'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=WaitForFirstConsumer' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-not-existing-vmclass-with-changing\n\nwaited for: 'jsonpath={.status.phase}=WaitForFirstConsumer'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:103 @ 10/11/25 17:19:18.666 < Exit [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/11/25 17:19:18.666 (16m42.142s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/11/25 17:19:18.666 The list of pods is empty; nothing to dump. < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/11/25 17:19:24.922 (6.256s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/11/25 17:19:24.922 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:126 @ 10/11/25 17:19:24.922 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:140 @ 10/11/25 17:19:24.922 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:152 @ 10/11/25 17:19:24.922 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:157 @ 10/11/25 17:19:24.923 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:163 @ 10/11/25 17:19:24.923 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:176 @ 10/11/25 17:19:24.923 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:181 @ 10/11/25 17:19:24.923 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:187 @ 10/11/25 17:19:24.923 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:202 @ 10/11/25 17:19:24.923 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:216 @ 10/11/25 17:19:24.923 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:240 @ 10/11/25 17:19:24.923 + + + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/11/25 17:19:24.923 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/11/25 17:19:26.902 (1.979s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/11/25 17:19:26.902 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/11/25 17:19:31.079 (4.177s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/11/25 17:19:31.079 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/11/25 17:19:31.079 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/11/25 17:19:31.08 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:97 @ 10/11/25 17:19:31.08 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/11/25 17:19:45.346 (14.266s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/11/25 17:19:45.346 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/11/25 17:19:45.346 (0s) + + + [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/11/25 17:20:26.987 This is the Progress Report generated when the suite timeout occurred: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 41.641s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [It] (Node Runtime: 41.641s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 At [By Step] VDs should be in Ready phase (Step Runtime: 41.641s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 Spec Goroutine goroutine 944 [sync.WaitGroup.Wait] sync.runtime_SemacquireWaitGroup(0x100140b30?, 0x60?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x140007215b0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000595340, 0x2, 0x101bb9e58?}, {0x1018d2204, 0x28}, {{0x0, 0x0, 0x0}, 0x14000653ef0, {0x14000d9db30, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1018d2204, 0x28}, {{0x0, 0x0, 0x0}, 0x14000653ef0, {0x14000d9db30, 0x25}, {0x14000498280, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1018d2204, 0x28}, {0x101890eef, 0x5}, {{0x0, 0x0, 0x0}, 0x14000653ef0, {0x14000d9db30, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 25 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 1060 [syscall] syscall.syscall6(0x100998040?, 0x12fc4cf80?, 0x1044b0108?, 0x90?, 0x140004dd008?, 0x14000eea120?, 0x140006c1a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140006c1a88?, 0x1001c26fc?, 0x90?, 0x1021be120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002f6380?, 0x140006c1ac4, 0x14000998040?, 0x140002f6310?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140008a8280) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140007a6488?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000176000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000176000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140004058e0?, 0x102255e88?, 0x140002f6000?}}, {0x102276488?, 0x140002f6000?}, {0x14000b2a000?, 0xc232b8bfaf5ea5e0?}, {0x10224ed80, 0x140008a8200}, {0x10224ed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140004058e0?, 0x103763728?, 0x103739fa0?}}, {0x102276488, 0x140002f6000}, {0x14000b2a000, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10227aa38?, 0x1400000f5a8?}, {0x101893fba?, 0x14000a5be88?}}, {0x1018d2204, 0x28}, {0x14000bcc480, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 944 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 1061 [syscall] syscall.syscall6(0x1007215f0?, 0x1046b9d38?, 0x1044b13e8?, 0x90?, 0x10373c480?, 0x14000146480?, 0x1400090ca58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400090ca88?, 0x1001c26fc?, 0x90?, 0x1021be120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002a8b60?, 0x1400090cac4, 0x140007215f0?, 0x140002a8a80?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140003e4fc0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000476908?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000197800) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000197800) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140004058e0?, 0x102255e88?, 0x140002a89a0?}}, {0x102276488?, 0x140002a89a0?}, {0x14000974600?, 0xc232b8bfaf5e2110?}, {0x10224ed80, 0x140003e4ec0}, {0x10224ed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140004058e0?, 0x103763728?, 0x103739fa0?}}, {0x102276488, 0x140002a89a0}, {0x14000974600, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10227aa38?, 0x1400000f5a8?}, {0x101893fba?, 0x14000a54688?}}, {0x1018d2204, 0x28}, {0x14000bcc4a0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 944 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) + > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/11/25 17:19:45.346 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 @ 10/11/25 17:19:45.346 [TIMEDOUT] A suite timeout occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/11/25 17:20:26.987 This is the Progress Report generated when the suite timeout occurred: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 41.641s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [It] (Node Runtime: 41.641s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 At [By Step] VDs should be in Ready phase (Step Runtime: 41.641s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 Spec Goroutine goroutine 944 [sync.WaitGroup.Wait] sync.runtime_SemacquireWaitGroup(0x100140b30?, 0x60?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x140007215b0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000595340, 0x2, 0x101bb9e58?}, {0x1018d2204, 0x28}, {{0x0, 0x0, 0x0}, 0x14000653ef0, {0x14000d9db30, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1018d2204, 0x28}, {{0x0, 0x0, 0x0}, 0x14000653ef0, {0x14000d9db30, 0x25}, {0x14000498280, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1018d2204, 0x28}, {0x101890eef, 0x5}, {{0x0, 0x0, 0x0}, 0x14000653ef0, {0x14000d9db30, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 25 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 1060 [syscall] syscall.syscall6(0x100998040?, 0x12fc4cf80?, 0x1044b0108?, 0x90?, 0x140004dd008?, 0x14000eea120?, 0x140006c1a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140006c1a88?, 0x1001c26fc?, 0x90?, 0x1021be120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002f6380?, 0x140006c1ac4, 0x14000998040?, 0x140002f6310?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140008a8280) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140007a6488?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000176000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000176000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140004058e0?, 0x102255e88?, 0x140002f6000?}}, {0x102276488?, 0x140002f6000?}, {0x14000b2a000?, 0xc232b8bfaf5ea5e0?}, {0x10224ed80, 0x140008a8200}, {0x10224ed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140004058e0?, 0x103763728?, 0x103739fa0?}}, {0x102276488, 0x140002f6000}, {0x14000b2a000, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10227aa38?, 0x1400000f5a8?}, {0x101893fba?, 0x14000a5be88?}}, {0x1018d2204, 0x28}, {0x14000bcc480, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 944 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 1061 [syscall] syscall.syscall6(0x1007215f0?, 0x1046b9d38?, 0x1044b13e8?, 0x90?, 0x10373c480?, 0x14000146480?, 0x1400090ca58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400090ca88?, 0x1001c26fc?, 0x90?, 0x1021be120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140002a8b60?, 0x1400090cac4, 0x140007215f0?, 0x140002a8a80?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140003e4fc0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000476908?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000197800) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000197800) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140004058e0?, 0x102255e88?, 0x140002a89a0?}}, {0x102276488?, 0x140002a89a0?}, {0x14000974600?, 0xc232b8bfaf5e2110?}, {0x10224ed80, 0x140003e4ec0}, {0x10224ed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140004058e0?, 0x103763728?, 0x103739fa0?}}, {0x102276488, 0x140002a89a0}, {0x14000974600, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10227aa38?, 0x1400000f5a8?}, {0x101893fba?, 0x14000a54688?}}, {0x1018d2204, 0x28}, {0x14000bcc4a0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 944 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/11/25 17:20:26.993 (41.647s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/11/25 17:20:26.993 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/11/25 17:20:32.633 (5.64s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x14000469008>: the container "virtualization-controller" was restarted: virtualization-controller-665cd6fcb-ndxzr the container "virtualization-controller" was restarted: virtualization-controller-665cd6fcb-rkvmw the container "virtualization-controller" was not found: virtualization-controller-744d54f474-5vrh9 { errs: [ <*errors.joinError | 0x14000468fa8>{ errs: [ <*errors.joinError | 0x14000468f48>{ errs: [ <*errors.errorString | 0x140007182a0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-665cd6fcb-ndxzr", }, ], }, <*errors.errorString | 0x140007182d0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-665cd6fcb-rkvmw", }, ], }, <*errors.errorString | 0x14000718300>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-744d54f474-5vrh9", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/11/25 17:20:33.642 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/11/25 17:20:32.635 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/11/25 17:20:32.635 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/11/25 17:20:32.635 [FAILED] Expected success, but got an error: <*errors.joinError | 0x14000469008>: the container "virtualization-controller" was restarted: virtualization-controller-665cd6fcb-ndxzr the container "virtualization-controller" was restarted: virtualization-controller-665cd6fcb-rkvmw the container "virtualization-controller" was not found: virtualization-controller-744d54f474-5vrh9 { errs: [ <*errors.joinError | 0x14000468fa8>{ errs: [ <*errors.joinError | 0x14000468f48>{ errs: [ <*errors.errorString | 0x140007182a0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-665cd6fcb-ndxzr", }, ], }, <*errors.errorString | 0x140007182d0>{ s: "the container \"virtualization-controller\" was restarted: virtualization-controller-665cd6fcb-rkvmw", }, ], }, <*errors.errorString | 0x14000718300>{ s: "the container \"virtualization-controller\" was not found: virtualization-controller-744d54f474-5vrh9", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:183 @ 10/11/25 17:20:33.642 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/11/25 17:20:33.642 (1.008s) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/11/25 17:20:33.642 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/11/25 17:20:49.76 (16.118s) + + + \ No newline at end of file diff --git a/artifacts/unique-cephrbd-20251012-153735-5317/junit.xml b/artifacts/unique-cephrbd-20251012-153735-5317/junit.xml new file mode 100644 index 0000000000..e31d6d833c --- /dev/null +++ b/artifacts/unique-cephrbd-20251012-153735-5317/junit.xml @@ -0,0 +1,782 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 16:00:10.195 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 16:00:14.917 (4.723s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 16:00:14.918 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 16:00:14.918 (0s) + + + + > Enter [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/12/25 16:00:14.92 [SKIPPED] Module SDN is disabled. Skipping all tests for module SDN. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:50 @ 10/12/25 16:00:15.059 < Exit [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/12/25 16:00:15.059 (139ms) > Enter [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/12/25 16:00:15.059 < Exit [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/12/25 16:00:15.059 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/12/25 16:00:15.059 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/12/25 16:00:15.059 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:111 @ 10/12/25 16:00:15.06 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:125 @ 10/12/25 16:00:15.06 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:141 @ 10/12/25 16:00:15.06 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:156 @ 10/12/25 16:00:15.06 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:169 @ 10/12/25 16:00:15.06 + + + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/12/25 16:00:15.06 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/12/25 16:00:17.071 (2.011s) > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 16:00:17.071 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 16:00:17.071 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/12/25 16:00:17.071 �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Failed to update resource", "err": "error updating status subresource: Put \"http://127.0.0.1:23915/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-disk-attachment/virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone/status\": context canceled", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-5073ae15-vd-root-automatic-with-hotplug-standalone", "namespace": "head-5073ae15-end-to-end-vm-disk-attachment", "reconcileID": "8820772b-8a8a-470d-8f16-f7fdd52ead1b", "time": "2025-10-12T13:00:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "error updating status subresource: Put \"http://127.0.0.1:23915/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-disk-attachment/virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone/status\": context canceled", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-5073ae15-vd-root-automatic-with-hotplug-standalone", "namespace": "head-5073ae15-end-to-end-vm-disk-attachment", "reconcileID": "8820772b-8a8a-470d-8f16-f7fdd52ead1b", "time": "2025-10-12T13:00:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "error updating status subresource: Put \"http://127.0.0.1:23915/apis/virtualization.deckhouse.io/v1alpha2/namespaces/head-5073ae15-end-to-end-vm-disk-attachment/virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone/status\": context canceled", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-5073ae15-vd-root-automatic-with-hotplug-standalone", "namespace": "head-5073ae15-end-to-end-vm-disk-attachment", "reconcileID": "8820772b-8a8a-470d-8f16-f7fdd52ead1b", "time": "2025-10-12T13:00:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Failed to update resource", "err": "error updating status subresource: context canceled", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-5073ae15-vd-attach-automatic-with-hotplug-standalone", "namespace": "head-5073ae15-end-to-end-vm-disk-attachment", "reconcileID": "9690f7b5-37db-4309-9ff1-03c1efc2ac39", "time": "2025-10-12T13:00:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Error occurred during reconciliation", "err": "error updating status subresource: context canceled", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-5073ae15-vd-attach-automatic-with-hotplug-standalone", "namespace": "head-5073ae15-end-to-end-vm-disk-attachment", "reconcileID": "9690f7b5-37db-4309-9ff1-03c1efc2ac39", "time": "2025-10-12T13:00:21Z" }�[0m �[31m�[1m this is the `Virtualization-controller` error! not the current `Ginkgo` context error: �[0m�[31m{ "level": "error", "msg": "Reconciler error", "err": "error updating status subresource: context canceled", "controller": "vd-controller", "handler": "", "ds": "", "collector": "", "name": "head-5073ae15-vd-attach-automatic-with-hotplug-standalone", "namespace": "head-5073ae15-end-to-end-vm-disk-attachment", "reconcileID": "9690f7b5-37db-4309-9ff1-03c1efc2ac39", "time": "2025-10-12T13:00:21Z" }�[0m < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/12/25 16:00:22.028 (4.957s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 16:00:22.028 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 16:00:22.028 (0s) + + + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 16:00:22.028 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 16:00:22.028 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/12/25 16:00:22.028 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:88 @ 10/12/25 16:00:22.028 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/12/25 16:00:51.985 (29.957s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 16:00:51.985 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 16:00:51.985 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-with-hotplug-standalone --namespace head-5073ae15-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/12/25 16:17:33.969 + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 16:00:51.986 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/12/25 16:00:51.986 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/12/25 16:00:51.986 STEP: VDs with consumers should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:99 @ 10/12/25 16:00:51.986 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-automatic-with-hotplug-standalone --namespace head-5073ae15-end-to-end-vm-disk-attachment --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-automatic-with-hotplug-standalone\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:100 @ 10/12/25 16:17:33.969 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/12/25 16:17:33.97 (16m41.982s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 16:17:33.97 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/12/25 16:17:40.144 (6.174s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/12/25 16:17:40.144 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/12/25 16:17:40.145 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/12/25 16:17:40.145 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/12/25 16:17:40.145 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/12/25 16:17:40.145 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/12/25 16:17:40.145 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/12/25 16:17:40.145 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/12/25 16:17:40.145 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/12/25 16:17:40.145 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/12/25 16:17:40.145 + + + > Enter [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/12/25 16:17:40.145 < Exit [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/12/25 16:17:42.105 (1.959s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/12/25 16:17:42.105 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/12/25 16:17:51.141 (9.036s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/12/25 16:17:51.141 < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/12/25 16:17:51.141 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:6, cap:6>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-a-not-b --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-a-not-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-d --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-d\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-affinity --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-affinity\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-selector --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-selector\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-b-not-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-b-not-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-c-and-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-c-and-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:103 @ 10/12/25 16:34:46.904 + > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/12/25 16:17:51.142 STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/12/25 16:17:51.142 END STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/12/25 16:18:02.11 (10.968s) STEP: `VirtualMachineClasses` should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:95 @ 10/12/25 16:18:02.11 END STEP: `VirtualMachineClasses` should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:95 @ 10/12/25 16:18:04.816 (2.706s) STEP: `VirtualDisks` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:102 @ 10/12/25 16:18:04.816 END STEP: `VirtualDisks` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:102 @ 10/12/25 16:34:46.904 (16m42.086s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:6, cap:6>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-a-not-b --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-a-not-b\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-d --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-d\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-affinity --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-affinity\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-node-selector --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-node-selector\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-b-not-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-b-not-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-c-and-a --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-c-and-a\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:103 @ 10/12/25 16:34:46.904 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/12/25 16:34:46.904 (16m55.761s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/12/25 16:34:46.904 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/12/25 16:34:53.028 (6.124s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:120 @ 10/12/25 16:34:53.028 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:283 @ 10/12/25 16:34:53.028 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:372 @ 10/12/25 16:34:53.028 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:467 @ 10/12/25 16:34:53.029 + + + > Enter [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/12/25 16:34:53.029 < Exit [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/12/25 16:34:54.992 (1.964s) > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 16:34:54.992 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 16:34:54.992 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/12/25 16:34:54.992 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/12/25 16:34:57.015 (2.023s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 16:34:57.016 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 16:34:57.016 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 16:34:57.016 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 16:34:57.016 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/12/25 16:34:57.016 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:68 @ 10/12/25 16:34:57.016 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/12/25 16:35:16.667 (19.651s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 16:35:16.667 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 16:35:16.667 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 16:35:16.668 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 16:35:16.668 (0s) > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/12/25 16:35:16.668 STEP: VM should be in Running phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:79 @ 10/12/25 16:35:16.668 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/12/25 16:35:19.613 (2.946s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 16:35:19.613 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 16:35:19.614 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 16:35:19.614 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 16:35:19.614 (0s) > Enter [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/12/25 16:35:19.614 < Exit [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/12/25 16:35:19.614 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 16:35:19.614 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 16:35:19.614 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 16:35:19.614 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 16:35:19.614 (0s) > Enter [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/12/25 16:35:19.614 < Exit [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/12/25 16:35:19.614 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 16:35:19.614 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 16:35:19.614 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 16:35:19.614 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/12/25 16:35:19.614 (0s) > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/12/25 16:35:19.614 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 16:35:19.614 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 16:36:29.95 (1m10.336s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/12/25 16:36:29.95 (1m10.336s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 16:36:29.95 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/12/25 16:36:29.95 (0s) + + + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/12/25 16:36:29.951 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/12/25 16:36:32.133 (2.183s) > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 16:36:32.133 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 16:36:32.133 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/12/25 16:36:32.133 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/12/25 16:36:34.996 (2.863s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 16:36:34.996 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 16:36:34.996 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 16:36:34.996 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 16:36:34.996 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/12/25 16:36:34.996 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:75 @ 10/12/25 16:36:34.996 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/12/25 16:36:40.785 (5.789s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 16:36:40.785 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 16:36:40.786 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-label-annotation --namespace head-5073ae15-end-to-end-vm-label-annotation --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1012 17:07:31.096100 80804 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-label-annotation\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 @ 10/12/25 17:16:53.565 + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 16:36:40.786 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 16:36:40.786 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/12/25 16:36:40.786 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 @ 10/12/25 16:36:40.786 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-label-annotation --namespace head-5073ae15-end-to-end-vm-label-annotation --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1012 17:07:31.096100 80804 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-label-annotation\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:87 @ 10/12/25 17:16:53.565 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/12/25 17:16:53.565 (16m42.124s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 17:16:53.565 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 17:16:59.602 (6.037s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/12/25 17:16:59.602 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/12/25 17:16:59.602 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/12/25 17:16:59.602 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/12/25 17:16:59.602 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/12/25 17:16:59.602 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/12/25 17:16:59.603 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/12/25 17:16:59.603 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/12/25 17:16:59.603 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/12/25 17:16:59.603 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/12/25 17:16:59.603 + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/12/25 17:17:01.573 + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/12/25 17:16:59.603 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/12/25 17:17:01.573 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/12/25 17:17:01.573 (1.97s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/12/25 17:17:01.573 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/12/25 17:17:07.562 (5.989s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:100 @ 10/12/25 17:17:07.562 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:111 @ 10/12/25 17:17:07.562 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:122 @ 10/12/25 17:17:07.562 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:133 @ 10/12/25 17:17:07.562 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:144 @ 10/12/25 17:17:07.563 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:164 @ 10/12/25 17:17:07.563 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:185 @ 10/12/25 17:17:07.563 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:229 @ 10/12/25 17:17:07.563 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:290 @ 10/12/25 17:17:07.563 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:341 @ 10/12/25 17:17:07.563 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:367 @ 10/12/25 17:17:07.563 + + + [FAILED] Warning: resource namespaces/head-5073ae15-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "ceph-pool-r2-csi-rbd" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x14000131840>: exit status 1 { ProcessState: { pid: 82071, status: 256, rusage: { Utime: { Sec: 0, Usec: 293824, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 63338, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 70664192, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5122, Majflt: 5, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 48, Msgrcv: 56, Nsignals: 361, Nvcsw: 342, Nivcsw: 2752, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/12/25 17:17:14.758 + > Enter [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/12/25 17:17:07.563 < Exit [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/12/25 17:17:10.574 (3.011s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/12/25 17:17:10.574 [FAILED] Warning: resource namespaces/head-5073ae15-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "ceph-pool-r2-csi-rbd" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x14000131840>: exit status 1 { ProcessState: { pid: 82071, status: 256, rusage: { Utime: { Sec: 0, Usec: 293824, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 63338, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 70664192, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5122, Majflt: 5, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 48, Msgrcv: 56, Nsignals: 361, Nvcsw: 342, Nivcsw: 2752, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/12/25 17:17:14.758 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/12/25 17:17:14.758 (4.184s) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/12/25 17:17:14.758 The list of pods is empty; nothing to dump. < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/12/25 17:17:20.599 (5.841s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:112 @ 10/12/25 17:17:20.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:147 @ 10/12/25 17:17:20.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:190 @ 10/12/25 17:17:20.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:196 @ 10/12/25 17:17:20.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:204 @ 10/12/25 17:17:20.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:224 @ 10/12/25 17:17:20.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:236 @ 10/12/25 17:17:20.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:258 @ 10/12/25 17:17:20.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:284 @ 10/12/25 17:17:20.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:294 @ 10/12/25 17:17:20.599 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:309 @ 10/12/25 17:17:20.6 + + + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/12/25 17:17:20.6 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/12/25 17:17:22.554 (1.954s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:62 @ 10/12/25 17:17:22.554 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:62 @ 10/12/25 17:17:51.566 (29.012s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 17:17:51.566 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 17:17:51.566 (0s) + + + > Enter [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/12/25 17:17:51.566 < Exit [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/12/25 17:17:53.012 (1.446s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 17:17:53.012 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 17:17:53.012 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/12/25 17:17:53.012 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:102 @ 10/12/25 17:17:53.012 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/12/25 17:18:00.454 (7.441s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 17:18:00.454 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 17:18:00.454 (0s) + + + > Enter [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/12/25 17:18:00.454 STEP: CVIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:113 @ 10/12/25 17:18:00.454 < Exit [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/12/25 17:18:03.392 (2.938s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 17:18:03.392 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 17:18:03.392 (0s) + + + > Enter [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/12/25 17:18:03.393 STEP: VMClasses should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:124 @ 10/12/25 17:18:03.393 < Exit [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/12/25 17:18:06.141 (2.748s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 17:18:06.141 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 17:18:06.141 (0s) + + + [FAILED] Timed out after 344.295s. Expected success, but got an error: <*errors.errorString | 0x140007fe050>: cannot patch VMIP "head-5073ae15-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip { s: "cannot patch VMIP \"head-5073ae15-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:138 @ 10/12/25 17:23:50.433 + > Enter [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/12/25 17:18:06.141 [FAILED] Timed out after 344.295s. Expected success, but got an error: <*errors.errorString | 0x140007fe050>: cannot patch VMIP "head-5073ae15-vm-custom-ip" with unnassigned IP address error: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip { s: "cannot patch VMIP \"head-5073ae15-vm-custom-ip\" with unnassigned IP address\nerror: timed out waiting for the condition on virtualmachineipaddresses/head-5073ae15-vm-custom-ip\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:138 @ 10/12/25 17:23:50.433 < Exit [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/12/25 17:23:50.434 (5m44.296s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 17:23:50.434 The list of pods is empty; nothing to dump. < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 17:23:57.781 (7.346s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:141 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:152 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:162 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:173 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:184 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:196 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:220 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:245 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:268 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:280 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:299 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:316 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:341 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:354 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:358 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:375 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:388 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:405 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:444 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:479 @ 10/12/25 17:23:57.781 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:497 @ 10/12/25 17:23:57.782 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:512 @ 10/12/25 17:23:57.782 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:528 @ 10/12/25 17:23:57.782 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:545 @ 10/12/25 17:23:57.782 + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.782 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:23:57.782 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.782 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.782 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:23:57.782 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.782 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.782 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:23:57.782 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.782 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.782 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:23:57.782 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.782 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.782 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:23:57.783 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.783 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.783 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:23:57.783 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.783 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.783 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:23:57.783 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.783 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.783 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:23:57.783 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.783 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.783 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:23:57.783 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:23:57.783 (0s) + + + > Enter [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/12/25 17:23:57.783 < Exit [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/12/25 17:23:59.729 (1.946s) > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 17:23:59.729 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 17:23:59.729 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/12/25 17:23:59.729 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/12/25 17:24:06.761 (7.032s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 17:24:06.761 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 17:24:06.761 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 17:24:06.761 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 17:24:06.761 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/12/25 17:24:06.761 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:91 @ 10/12/25 17:24:06.761 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/12/25 17:24:11.625 (4.864s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 17:24:11.625 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 17:24:11.625 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-not-existing-vmclass-with-changing\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:103 @ 10/12/25 17:40:53.774 + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 17:24:11.625 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/12/25 17:24:11.625 (0s) > Enter [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/12/25 17:24:11.625 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:102 @ 10/12/25 17:24:11.625 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-root-not-existing-vmclass-with-changing --namespace head-5073ae15-end-to-end-sizing-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-root-not-existing-vmclass-with-changing\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:103 @ 10/12/25 17:40:53.774 < Exit [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/12/25 17:40:53.775 (16m42.158s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 17:40:53.775 The list of pods is empty; nothing to dump. < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/12/25 17:41:00.164 (6.389s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/12/25 17:41:00.164 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:126 @ 10/12/25 17:41:00.164 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:140 @ 10/12/25 17:41:00.164 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:152 @ 10/12/25 17:41:00.164 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:157 @ 10/12/25 17:41:00.164 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:163 @ 10/12/25 17:41:00.165 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:176 @ 10/12/25 17:41:00.165 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:181 @ 10/12/25 17:41:00.165 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:187 @ 10/12/25 17:41:00.165 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:202 @ 10/12/25 17:41:00.165 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:216 @ 10/12/25 17:41:00.165 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:240 @ 10/12/25 17:41:00.165 + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1012 18:25:50.079961 84017 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nW1012 19:14:48.506065 84017 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nW1012 19:46:49.136743 84017 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1012 18:25:56.831828 84018 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nW1012 19:32:35.365049 84018 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 @ 10/12/25 19:47:49.186 + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/12/25 17:41:00.165 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/12/25 17:41:02.137 (1.972s) > Enter [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/12/25 17:41:02.137 < Exit [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/12/25 17:41:07.038 (4.901s) > Enter [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/12/25 17:41:07.038 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 @ 10/12/25 17:41:07.038 [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1012 18:25:50.079961 84017 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nW1012 19:14:48.506065 84017 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nW1012 19:46:49.136743 84017 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1012 18:25:56.831828 84018 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nW1012 19:32:35.365049 84018 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 @ 10/12/25 19:47:49.186 < Exit [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/12/25 19:47:49.186 (16m42.102s) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/12/25 19:47:49.186 The list of pods is empty; nothing to dump. STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 19:47:55.404 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 19:48:00.681 (5.277s) < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/12/25 19:48:00.681 (11.495s) + + + + > Enter [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/12/25 19:48:00.682 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:41 @ 10/12/25 19:48:00.682 < Exit [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/12/25 19:48:00.682 (0s) > Enter [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/12/25 19:48:00.682 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 19:48:00.682 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 19:48:05.66 (4.978s) < Exit [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/12/25 19:48:05.66 (4.978s) + + + + > Enter [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/12/25 19:48:05.66 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:73 @ 10/12/25 19:48:05.66 < Exit [BeforeAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:69 @ 10/12/25 19:48:05.66 (0s) > Enter [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/12/25 19:48:05.66 < Exit [AfterEach] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_live_migration_tcp_session_test.go:79 @ 10/12/25 19:48:05.66 (0s) > Enter [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/12/25 19:48:05.66 < Exit [AfterAll] VirtualMachineLiveMigrationTCPSession - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/framework/framework.go:76 @ 10/12/25 19:48:05.66 (0s) + + + > Enter [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/12/25 19:48:05.661 < Exit [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/12/25 19:48:05.662 (2ms) > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 19:48:05.662 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 19:48:05.662 (0s) > Enter [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/12/25 19:48:05.662 < Exit [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/12/25 19:48:07.297 (1.635s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 19:48:07.297 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 19:48:07.297 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 19:48:07.298 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 19:48:07.298 (0s) > Enter [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/12/25 19:48:07.298 STEP: Project should be deployed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:73 @ 10/12/25 19:48:07.298 < Exit [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/12/25 19:48:10.044 (2.746s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 19:48:10.044 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 19:48:10.044 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 19:48:10.044 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 19:48:10.044 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/12/25 19:48:10.044 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/12/25 19:48:13.127 (3.082s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 19:48:13.127 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 19:48:13.127 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 19:48:13.127 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 19:48:13.127 (0s) > Enter [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/12/25 19:48:13.127 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/12/25 19:48:13.127 < Exit [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/12/25 19:48:19.793 (6.666s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 19:48:19.793 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 19:48:19.793 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-cvi --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-cvi\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/12/25 20:05:02.281 + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 19:48:19.793 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 19:48:19.793 (0s) > Enter [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/12/25 19:48:19.793 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/12/25 19:48:19.793 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualdisks.virtualization.deckhouse.io head-5073ae15-vd-cvi --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualdisks/head-5073ae15-vd-cvi\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/12/25 20:05:02.281 < Exit [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/12/25 20:05:02.281 (16m42.487s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 20:05:02.281 The list of pods is empty; nothing to dump. < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 20:05:09.44 (7.159s) > Enter [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/12/25 20:05:09.44 STEP: Delete manifests - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:50 @ 10/12/25 20:05:09.44 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 20:05:09.44 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 20:05:11.414 (1.974s) < Exit [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/12/25 20:05:11.414 (1.975s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/12/25 20:05:11.415 + + + > Enter [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/12/25 20:05:11.415 < Exit [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/12/25 20:05:14.598 (3.183s) > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/12/25 20:05:14.598 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/12/25 20:05:14.598 (0s) > Enter [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/12/25 20:05:14.598 STEP: Create a vmip automatically and check its binding with a lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:81 @ 10/12/25 20:05:14.598 STEP: Remove label from the lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:87 @ 10/12/25 20:05:15.63 STEP: Wait for the label to be restored by the controller - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:96 @ 10/12/25 20:05:15.75 < Exit [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/12/25 20:05:15.992 (1.393s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/12/25 20:05:15.992 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/12/25 20:05:15.992 (0s) + + + > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/12/25 20:05:15.993 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/12/25 20:05:15.993 (0s) > Enter [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/12/25 20:05:15.993 STEP: Create an intermediate vmip automatically to allocate a new ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:109 @ 10/12/25 20:05:15.993 STEP: Delete the intermediate vmip automatically and check that the lease is released - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:115 @ 10/12/25 20:05:16.351 STEP: Reuse the released lease with a static vmip - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:124 @ 10/12/25 20:05:16.592 STEP: Delete the static vmip and lease, then create another static vmip with this ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:132 @ 10/12/25 20:05:17.066 < Exit [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/12/25 20:05:17.695 (1.702s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/12/25 20:05:17.695 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/12/25 20:05:17.695 (0s) + + + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/12/25 20:05:17.695 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/12/25 20:05:19.694 (1.999s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/12/25 20:05:19.694 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/12/25 20:05:23.805 (4.111s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 20:05:23.805 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 20:05:23.805 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/12/25 20:05:23.805 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:97 @ 10/12/25 20:05:23.805 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/12/25 20:05:39.213 (15.408s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 20:05:39.213 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 20:05:39.213 (0s) + + + [INTERRUPTED] Interrupted by User In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/12/25 20:17:24.214 This is the Progress Report generated when the interrupt was received: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 11m45s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [It] (Node Runtime: 11m45s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 At [By Step] VDs should be in Ready phase (Step Runtime: 11m45s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 Spec Goroutine goroutine 805 [sync.WaitGroup.Wait, 12 minutes] sync.runtime_SemacquireWaitGroup(0x100504b30?, 0xa0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000bfa010) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000be2000, 0x2, 0x101f7de58?}, {0x101c96204, 0x28}, {{0x0, 0x0, 0x0}, 0x140006cd710, {0x140009ed110, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x101c96204, 0x28}, {{0x0, 0x0, 0x0}, 0x140006cd710, {0x140009ed110, 0x25}, {0x140009eea80, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x101c96204, 0x28}, {0x101c54eef, 0x5}, {{0x0, 0x0, 0x0}, 0x140006cd710, {0x140009ed110, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 11 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 761 [syscall, 12 minutes] syscall.syscall6(0x100da6a10?, 0x104bd2228?, 0x1048745c0?, 0x90?, 0x14000580808?, 0x14000a8e360?, 0x14000b33a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000b33a88?, 0x1005866fc?, 0x90?, 0x102582120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029b110?, 0x14000b33ac4, 0x14000da6a10?, 0x1400029b0a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140002413c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000db6408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140005b4300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140005b4300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400017f9e0?, 0x102619e88?, 0x1400029afc0?}}, {0x10263a488?, 0x1400029afc0?}, {0x14000028780?, 0xc23316d81b4235e0?}, {0x102612d80, 0x14000241300}, {0x102612d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400017f9e0?, 0x103b27728?, 0x103afdfa0?}}, {0x10263a488, 0x1400029afc0}, {0x14000028780, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10263ea38?, 0x14000496a38?}, {0x101c57fba?, 0x14000fb4e88?}}, {0x101c96204, 0x28}, {0x140009fc1c0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 805 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 762 [syscall, 12 minutes] syscall.syscall6(0x100bfa050?, 0x10fe07680?, 0x104874f30?, 0x90?, 0x14000980008?, 0x14000662120?, 0x14000b35a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000b35a88?, 0x1005866fc?, 0x90?, 0x102582120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400027e930?, 0x14000b35ac4, 0x14000bfa050?, 0x1400027e850?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000674300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140005ca408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140002cc000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140002cc000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400017f9e0?, 0x102619e88?, 0x140002c0000?}}, {0x10263a488?, 0x140002c0000?}, {0x140006fe000?, 0xc23316d81b4264c0?}, {0x102612d80, 0x14000bc8080}, {0x102612d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400017f9e0?, 0x103b27728?, 0x103afdfa0?}}, {0x10263a488, 0x140002c0000}, {0x140006fe000, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10263ea38?, 0x14000496a38?}, {0x101c57fba?, 0x140009dd688?}}, {0x101c96204, 0x28}, {0x140009fc1e0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 805 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) Other Goroutines goroutine 11 [running] github.com/onsi/ginkgo/v2/internal.extractRunningGoroutines() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:181 github.com/onsi/ginkgo/v2/internal.NewProgressReport(_, {{0x140005b2500, 0x2, 0x2}, {0x1400029abd0, 0x2, 0x2}, {0x14000893410, 0x2, 0x2}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:75 github.com/onsi/ginkgo/v2/internal.(*Suite).generateProgressReport(_, _) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:381 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode(_, {0xdc, 0x4, {0x101c68e1e, 0x11}, 0x1400017e850, {{0x102d418e1, 0x55}, 0x6b, {0x0, ...}, ...}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:980 github.com/onsi/ginkgo/v2/internal.(*group).attemptSpec(0x14000dd57f0, 0x1, {{0x14000677208?, 0x1400029abd0?, 0x2?}, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:199 github.com/onsi/ginkgo/v2/internal.(*group).run(0x14000dd57f0, {0x14000d32008?, 0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:349 github.com/onsi/ginkgo/v2/internal.(*Suite).runSpecs(0x140004b3508, {0x101c54f2b, 0x5}, {0x103b27728, 0x0, 0x0}, {0x1400005a0f4, 0x3d}, 0x0, {0x14000092c08, ...}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:489 github.com/onsi/ginkgo/v2/internal.(*Suite).Run(_, {_, _}, {_, _, _}, {_, _}, _, {_, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:130 github.com/onsi/ginkgo/v2.RunSpecs({0x1026120a0, 0x140005836c0}, {0x101c54f2b, 0x5}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:300 github.com/deckhouse/virtualization/tests/e2e.TestE2E(0x140005836c0) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:76 testing.tRunner(0x140005836c0, 0x102601c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.(*T).Run in goroutine 1 /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1997 goroutine 1 [chan receive, 123 minutes] testing.(*T).Run(0x14000583500, {0x101c5799a?, 0x14000379b38?}, 0x102601c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2005 testing.runTests.func1(0x14000583500) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2477 testing.tRunner(0x14000583500, 0x14000379c68) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.runTests(0x140004966f0, {0x103aacfd0, 0x1, 0x1}, {0x140005b3100?, 0xf?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2475 testing.(*M).Run(0x1400025cc80) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2337 main.main() _testmain.go:45 goroutine 41 [syscall] os/signal.signal_recv() /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sigqueue.go:149 os/signal.loop() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal_unix.go:23 os/signal.Notify.func1.1 in goroutine 11 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal.go:152 goroutine 66 [select] github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts.func2(0x140006e8000?) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:131 github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts in goroutine 11 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:128 goroutine 82 [select, 123 minutes] github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal.func1() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:32 github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal in goroutine 11 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:30 goroutine 811 [select, 12 minutes] os/exec.(*Cmd).watchCtx(0x140005b4300, 0x1400070cee0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 761 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 800 [IO wait, 12 minutes] internal/poll.runtime_pollWait(0x104b73a00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000caa3c0?, 0x1400094a000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000caa3c0, {0x1400094a000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140005a6040, {0x1400094a000?, 0x14000fba5a8?, 0x1004c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102612d80, 0x14000bc80c0}, {0x102611ec8, 0x140005a6068}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000496a01?, {0x102612d80, 0x14000bc80c0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x103aad230?, {0x102612d80?, 0x14000bc80c0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102612d80, 0x14000bc80c0}, {0x102611fc0, 0x140005a6040}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140008e8008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 762 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 801 [select, 12 minutes] os/exec.(*Cmd).watchCtx(0x140002cc000, 0x140001180e0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 762 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 809 [IO wait, 12 minutes] internal/poll.runtime_pollWait(0x104b74400, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000694ba0?, 0x14000940000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000694ba0, {0x14000940000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140000a2288, {0x14000940000?, 0x14000fb5da8?, 0x1004c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102612d80, 0x14000241300}, {0x102611ec8, 0x140005a6060}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000fb5e01?, {0x102612d80, 0x14000241300}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x103aad230?, {0x102612d80?, 0x14000241300?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102612d80, 0x14000241300}, {0x102611fc0, 0x140000a2288}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400070c620?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 761 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 799 [IO wait, 12 minutes] internal/poll.runtime_pollWait(0x104b73e00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000caa300?, 0x140009f0000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000caa300, {0x140009f0000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140005a6028, {0x140009f0000?, 0x14000fb55a8?, 0x1004c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102612d80, 0x14000bc8080}, {0x102611ec8, 0x14000702058}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1400070cd01?, {0x102612d80, 0x14000bc8080}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x103aad230?, {0x102612d80?, 0x14000bc8080?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102612d80, 0x14000bc8080}, {0x102611fc0, 0x140005a6028}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140005b4180?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 762 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 810 [IO wait, 12 minutes] internal/poll.runtime_pollWait(0x104b74e00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000694d20?, 0x140009aa000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000694d20, {0x140009aa000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140000a22b8, {0x140009aa000?, 0x14000fb4da8?, 0x1004c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102612d80, 0x14000241380}, {0x102611ec8, 0x14000702050}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000496a01?, {0x102612d80, 0x14000241380}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x103aad230?, {0x102612d80?, 0x14000241380?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102612d80, 0x14000241380}, {0x102611fc0, 0x140000a22b8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x14000bfa008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 761 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/12/25 20:05:39.213 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 @ 10/12/25 20:05:39.213 [INTERRUPTED] Interrupted by User In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/12/25 20:17:24.214 This is the Progress Report generated when the interrupt was received: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 11m45s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [It] (Node Runtime: 11m45s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 At [By Step] VDs should be in Ready phase (Step Runtime: 11m45s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 Spec Goroutine goroutine 805 [sync.WaitGroup.Wait, 12 minutes] sync.runtime_SemacquireWaitGroup(0x100504b30?, 0xa0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000bfa010) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000be2000, 0x2, 0x101f7de58?}, {0x101c96204, 0x28}, {{0x0, 0x0, 0x0}, 0x140006cd710, {0x140009ed110, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x101c96204, 0x28}, {{0x0, 0x0, 0x0}, 0x140006cd710, {0x140009ed110, 0x25}, {0x140009eea80, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x101c96204, 0x28}, {0x101c54eef, 0x5}, {{0x0, 0x0, 0x0}, 0x140006cd710, {0x140009ed110, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 11 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 761 [syscall, 12 minutes] syscall.syscall6(0x100da6a10?, 0x104bd2228?, 0x1048745c0?, 0x90?, 0x14000580808?, 0x14000a8e360?, 0x14000b33a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000b33a88?, 0x1005866fc?, 0x90?, 0x102582120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029b110?, 0x14000b33ac4, 0x14000da6a10?, 0x1400029b0a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140002413c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000db6408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140005b4300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140005b4300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400017f9e0?, 0x102619e88?, 0x1400029afc0?}}, {0x10263a488?, 0x1400029afc0?}, {0x14000028780?, 0xc23316d81b4235e0?}, {0x102612d80, 0x14000241300}, {0x102612d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400017f9e0?, 0x103b27728?, 0x103afdfa0?}}, {0x10263a488, 0x1400029afc0}, {0x14000028780, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10263ea38?, 0x14000496a38?}, {0x101c57fba?, 0x14000fb4e88?}}, {0x101c96204, 0x28}, {0x140009fc1c0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 805 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 762 [syscall, 12 minutes] syscall.syscall6(0x100bfa050?, 0x10fe07680?, 0x104874f30?, 0x90?, 0x14000980008?, 0x14000662120?, 0x14000b35a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000b35a88?, 0x1005866fc?, 0x90?, 0x102582120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400027e930?, 0x14000b35ac4, 0x14000bfa050?, 0x1400027e850?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000674300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140005ca408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140002cc000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140002cc000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400017f9e0?, 0x102619e88?, 0x140002c0000?}}, {0x10263a488?, 0x140002c0000?}, {0x140006fe000?, 0xc23316d81b4264c0?}, {0x102612d80, 0x14000bc8080}, {0x102612d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400017f9e0?, 0x103b27728?, 0x103afdfa0?}}, {0x10263a488, 0x140002c0000}, {0x140006fe000, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10263ea38?, 0x14000496a38?}, {0x101c57fba?, 0x140009dd688?}}, {0x101c96204, 0x28}, {0x140009fc1e0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 805 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) Other Goroutines goroutine 11 [running] github.com/onsi/ginkgo/v2/internal.extractRunningGoroutines() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:181 github.com/onsi/ginkgo/v2/internal.NewProgressReport(_, {{0x140005b2500, 0x2, 0x2}, {0x1400029abd0, 0x2, 0x2}, {0x14000893410, 0x2, 0x2}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:75 github.com/onsi/ginkgo/v2/internal.(*Suite).generateProgressReport(_, _) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:381 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode(_, {0xdc, 0x4, {0x101c68e1e, 0x11}, 0x1400017e850, {{0x102d418e1, 0x55}, 0x6b, {0x0, ...}, ...}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:980 github.com/onsi/ginkgo/v2/internal.(*group).attemptSpec(0x14000dd57f0, 0x1, {{0x14000677208?, 0x1400029abd0?, 0x2?}, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:199 github.com/onsi/ginkgo/v2/internal.(*group).run(0x14000dd57f0, {0x14000d32008?, 0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:349 github.com/onsi/ginkgo/v2/internal.(*Suite).runSpecs(0x140004b3508, {0x101c54f2b, 0x5}, {0x103b27728, 0x0, 0x0}, {0x1400005a0f4, 0x3d}, 0x0, {0x14000092c08, ...}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:489 github.com/onsi/ginkgo/v2/internal.(*Suite).Run(_, {_, _}, {_, _, _}, {_, _}, _, {_, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:130 github.com/onsi/ginkgo/v2.RunSpecs({0x1026120a0, 0x140005836c0}, {0x101c54f2b, 0x5}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:300 github.com/deckhouse/virtualization/tests/e2e.TestE2E(0x140005836c0) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:76 testing.tRunner(0x140005836c0, 0x102601c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.(*T).Run in goroutine 1 /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1997 goroutine 1 [chan receive, 123 minutes] testing.(*T).Run(0x14000583500, {0x101c5799a?, 0x14000379b38?}, 0x102601c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2005 testing.runTests.func1(0x14000583500) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2477 testing.tRunner(0x14000583500, 0x14000379c68) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.runTests(0x140004966f0, {0x103aacfd0, 0x1, 0x1}, {0x140005b3100?, 0xf?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2475 testing.(*M).Run(0x1400025cc80) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2337 main.main() _testmain.go:45 goroutine 41 [syscall] os/signal.signal_recv() /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sigqueue.go:149 os/signal.loop() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal_unix.go:23 os/signal.Notify.func1.1 in goroutine 11 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal.go:152 goroutine 66 [select] github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts.func2(0x140006e8000?) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:131 github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts in goroutine 11 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:128 goroutine 82 [select, 123 minutes] github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal.func1() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:32 github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal in goroutine 11 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:30 goroutine 811 [select, 12 minutes] os/exec.(*Cmd).watchCtx(0x140005b4300, 0x1400070cee0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 761 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 800 [IO wait, 12 minutes] internal/poll.runtime_pollWait(0x104b73a00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000caa3c0?, 0x1400094a000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000caa3c0, {0x1400094a000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140005a6040, {0x1400094a000?, 0x14000fba5a8?, 0x1004c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102612d80, 0x14000bc80c0}, {0x102611ec8, 0x140005a6068}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000496a01?, {0x102612d80, 0x14000bc80c0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x103aad230?, {0x102612d80?, 0x14000bc80c0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102612d80, 0x14000bc80c0}, {0x102611fc0, 0x140005a6040}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140008e8008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 762 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 801 [select, 12 minutes] os/exec.(*Cmd).watchCtx(0x140002cc000, 0x140001180e0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 762 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 809 [IO wait, 12 minutes] internal/poll.runtime_pollWait(0x104b74400, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000694ba0?, 0x14000940000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000694ba0, {0x14000940000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140000a2288, {0x14000940000?, 0x14000fb5da8?, 0x1004c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102612d80, 0x14000241300}, {0x102611ec8, 0x140005a6060}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000fb5e01?, {0x102612d80, 0x14000241300}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x103aad230?, {0x102612d80?, 0x14000241300?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102612d80, 0x14000241300}, {0x102611fc0, 0x140000a2288}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400070c620?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 761 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 799 [IO wait, 12 minutes] internal/poll.runtime_pollWait(0x104b73e00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000caa300?, 0x140009f0000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000caa300, {0x140009f0000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140005a6028, {0x140009f0000?, 0x14000fb55a8?, 0x1004c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102612d80, 0x14000bc8080}, {0x102611ec8, 0x14000702058}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1400070cd01?, {0x102612d80, 0x14000bc8080}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x103aad230?, {0x102612d80?, 0x14000bc8080?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102612d80, 0x14000bc8080}, {0x102611fc0, 0x140005a6028}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140005b4180?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 762 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 810 [IO wait, 12 minutes] internal/poll.runtime_pollWait(0x104b74e00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000694d20?, 0x140009aa000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000694d20, {0x140009aa000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140000a22b8, {0x140009aa000?, 0x14000fb4da8?, 0x1004c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102612d80, 0x14000241380}, {0x102611ec8, 0x14000702050}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000496a01?, {0x102612d80, 0x14000241380}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x103aad230?, {0x102612d80?, 0x14000241380?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102612d80, 0x14000241380}, {0x102611fc0, 0x140000a22b8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x14000bfa008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 761 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 Interrupted by User First interrupt received; Ginkgo will run any cleanup and reporting nodes but will skip all remaining specs. Interrupt again to skip cleanup. Here's a current progress report: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 11m45s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [It] (Node Runtime: 11m45s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 At [By Step] VDs should be in Ready phase (Step Runtime: 11m45s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 Spec Goroutine goroutine 805 [sync.WaitGroup.Wait, 12 minutes] sync.runtime_SemacquireWaitGroup(0x100504b30?, 0xa0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000bfa010) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000be2000, 0x2, 0x101f7de58?}, {0x101c96204, 0x28}, {{0x0, 0x0, 0x0}, 0x140006cd710, {0x140009ed110, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x101c96204, 0x28}, {{0x0, 0x0, 0x0}, 0x140006cd710, {0x140009ed110, 0x25}, {0x140009eea80, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x101c96204, 0x28}, {0x101c54eef, 0x5}, {{0x0, 0x0, 0x0}, 0x140006cd710, {0x140009ed110, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 11 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 761 [syscall, 12 minutes] syscall.syscall6(0x100da6a10?, 0x104bd2228?, 0x1048745c0?, 0x90?, 0x14000580808?, 0x14000a8e360?, 0x14000b33a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000b33a88?, 0x1005866fc?, 0x90?, 0x102582120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029b110?, 0x14000b33ac4, 0x14000da6a10?, 0x1400029b0a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140002413c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000db6408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140005b4300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140005b4300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400017f9e0?, 0x102619e88?, 0x1400029afc0?}}, {0x10263a488?, 0x1400029afc0?}, {0x14000028780?, 0xc23316d81b4235e0?}, {0x102612d80, 0x14000241300}, {0x102612d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400017f9e0?, 0x103b27728?, 0x103afdfa0?}}, {0x10263a488, 0x1400029afc0}, {0x14000028780, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10263ea38?, 0x14000496a38?}, {0x101c57fba?, 0x14000fb4e88?}}, {0x101c96204, 0x28}, {0x140009fc1c0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 805 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 762 [syscall, 12 minutes] syscall.syscall6(0x100bfa050?, 0x10fe07680?, 0x104874f30?, 0x90?, 0x14000980008?, 0x14000662120?, 0x14000b35a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000b35a88?, 0x1005866fc?, 0x90?, 0x102582120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400027e930?, 0x14000b35ac4, 0x14000bfa050?, 0x1400027e850?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000674300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140005ca408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140002cc000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140002cc000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400017f9e0?, 0x102619e88?, 0x140002c0000?}}, {0x10263a488?, 0x140002c0000?}, {0x140006fe000?, 0xc23316d81b4264c0?}, {0x102612d80, 0x14000bc8080}, {0x102612d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400017f9e0?, 0x103b27728?, 0x103afdfa0?}}, {0x10263a488, 0x140002c0000}, {0x140006fe000, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10263ea38?, 0x14000496a38?}, {0x101c57fba?, 0x140009dd688?}}, {0x101c96204, 0x28}, {0x140009fc1e0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 805 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/12/25 20:17:24.216 (11m45.002s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 20:17:24.216 [INTERRUPTED] Interrupted by User In [AfterEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 20:17:27.092 This is the Progress Report generated when the interrupt was received: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 11m47.878s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [AfterEach] (Node Runtime: 2.876s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 Spec Goroutine goroutine 814 [syscall] syscall.syscall6(0x100da7c00?, 0x104bd2228?, 0x1048745c0?, 0x90?, 0x14000680008?, 0x14000a8e6c0?, 0x1400036cb08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400036cb38?, 0x1005866fc?, 0x90?, 0x102582120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029a850?, 0x1400036cb74, 0x14000da7c00?, 0x1400029a7e0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000241200) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000db6c08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140005b4000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140005b4000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400017f9e0?, 0x102619e88?, 0x1400029a690?}}, {0x10263a488?, 0x1400029a690?}, {0x14000aca4d0?, 0xc23316908cec3e28?}, {0x102612d80, 0x14000241100}, {0x102612d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400017f9e0?, 0x103b27728?, 0x103afdfa0?}}, {0x10263a488, 0x1400029a690}, {0x14000aca4d0, 0xb0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.Get({{0x10263ea38?, 0x14000496a38?}, {0x101c57fba?, 0x140009fc9c0?}}, {0x101caf2e2, 0x31}, {{0x0, 0x0, 0x0}, 0x0, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:208 | ctx, cancel := context.WithTimeout(context.Background(), MediumTimeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.SaveTestCaseResources(0x140006cd710, {0x140008997e8, 0x11}, {0x140009ed110, 0x25}, {0x101c54370, 0x4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:765 | errorFileName := fmt.Sprintf("%s/e2e_failed__%s__%s_error.txt", dumpPath, labels["testcase"], additional) | > cmdr := kubectl.Get( | "virtualization,cvi,vmc,intvirt,pod,volumesnapshot", | kc.GetOptions{ > github.com/deckhouse/virtualization/tests/e2e.SaveTestCaseDump(0x140006cd710, {0x101c68e1e, 0x11}, {0x140009ed110, 0x25}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:757 | } | > SaveTestCaseResources(labels, additional, namespace, tmpDir) | SavePodLogsAndDescriptions(labels, additional, namespace, tmpDir) | } > github.com/deckhouse/virtualization/tests/e2e.init.func12.2() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:68 | AfterEach(func() { | if CurrentSpecReport().Failed() { > SaveTestCaseDump(testCaseLabel, CurrentSpecReport().LeafNodeText, ns) | } | }) github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 11 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 761 [syscall, 12 minutes] syscall.syscall6(0x100da6a10?, 0x104bd2228?, 0x1048745c0?, 0x90?, 0x14000580808?, 0x14000a8e360?, 0x14000b33a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000b33a88?, 0x1005866fc?, 0x90?, 0x102582120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029b110?, 0x14000b33ac4, 0x14000da6a10?, 0x1400029b0a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140002413c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000db6408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140005b4300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140005b4300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400017f9e0?, 0x102619e88?, 0x1400029afc0?}}, {0x10263a488?, 0x1400029afc0?}, {0x14000028780?, 0xc23316d81b4235e0?}, {0x102612d80, 0x14000241300}, {0x102612d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400017f9e0?, 0x103b27728?, 0x103afdfa0?}}, {0x10263a488, 0x1400029afc0}, {0x14000028780, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10263ea38?, 0x14000496a38?}, {0x101c57fba?, 0x14000fb4e88?}}, {0x101c96204, 0x28}, {0x140009fc1c0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 805 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 805 [sync.WaitGroup.Wait, 12 minutes] sync.runtime_SemacquireWaitGroup(0x100504b30?, 0xa0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000bfa010) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000be2000, 0x2, 0x101f7de58?}, {0x101c96204, 0x28}, {{0x0, 0x0, 0x0}, 0x140006cd710, {0x140009ed110, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x101c96204, 0x28}, {{0x0, 0x0, 0x0}, 0x140006cd710, {0x140009ed110, 0x25}, {0x140009eea80, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x101c96204, 0x28}, {0x101c54eef, 0x5}, {{0x0, 0x0, 0x0}, 0x140006cd710, {0x140009ed110, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 11 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 goroutine 762 [syscall, 12 minutes] syscall.syscall6(0x100bfa050?, 0x10fe07680?, 0x104874f30?, 0x90?, 0x14000980008?, 0x14000662120?, 0x14000b35a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000b35a88?, 0x1005866fc?, 0x90?, 0x102582120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400027e930?, 0x14000b35ac4, 0x14000bfa050?, 0x1400027e850?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000674300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140005ca408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140002cc000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140002cc000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400017f9e0?, 0x102619e88?, 0x140002c0000?}}, {0x10263a488?, 0x140002c0000?}, {0x140006fe000?, 0xc23316d81b4264c0?}, {0x102612d80, 0x14000bc8080}, {0x102612d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400017f9e0?, 0x103b27728?, 0x103afdfa0?}}, {0x10263a488, 0x140002c0000}, {0x140006fe000, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10263ea38?, 0x14000496a38?}, {0x101c57fba?, 0x140009dd688?}}, {0x101c96204, 0x28}, {0x140009fc1e0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 805 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) Other Goroutines goroutine 11 [running] github.com/onsi/ginkgo/v2/internal.extractRunningGoroutines() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:181 github.com/onsi/ginkgo/v2/internal.NewProgressReport(_, {{0x140005b2500, 0x2, 0x2}, {0x1400029abd0, 0x2, 0x2}, {0x14000893410, 0x2, 0x2}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:75 github.com/onsi/ginkgo/v2/internal.(*Suite).generateProgressReport(_, _) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:381 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode(_, {0xd6, 0x20, {0x0, 0x0}, 0x1400017e650, {{0x102d418e1, 0x55}, 0x42, {0x0, ...}, ...}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:980 github.com/onsi/ginkgo/v2/internal.(*group).attemptSpec(0x14000dd57f0, 0x1, {{0x14000677208?, 0x1400029abd0?, 0x2?}, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:286 github.com/onsi/ginkgo/v2/internal.(*group).run(0x14000dd57f0, {0x14000d32008?, 0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:349 github.com/onsi/ginkgo/v2/internal.(*Suite).runSpecs(0x140004b3508, {0x101c54f2b, 0x5}, {0x103b27728, 0x0, 0x0}, {0x1400005a0f4, 0x3d}, 0x0, {0x14000092c08, ...}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:489 github.com/onsi/ginkgo/v2/internal.(*Suite).Run(_, {_, _}, {_, _, _}, {_, _}, _, {_, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:130 github.com/onsi/ginkgo/v2.RunSpecs({0x1026120a0, 0x140005836c0}, {0x101c54f2b, 0x5}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:300 github.com/deckhouse/virtualization/tests/e2e.TestE2E(0x140005836c0) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:76 testing.tRunner(0x140005836c0, 0x102601c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.(*T).Run in goroutine 1 /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1997 goroutine 1 [chan receive, 123 minutes] testing.(*T).Run(0x14000583500, {0x101c5799a?, 0x14000379b38?}, 0x102601c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2005 testing.runTests.func1(0x14000583500) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2477 testing.tRunner(0x14000583500, 0x14000379c68) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.runTests(0x140004966f0, {0x103aacfd0, 0x1, 0x1}, {0x140005b3100?, 0xf?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2475 testing.(*M).Run(0x1400025cc80) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2337 main.main() _testmain.go:45 goroutine 41 [syscall] os/signal.signal_recv() /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sigqueue.go:149 os/signal.loop() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal_unix.go:23 os/signal.Notify.func1.1 in goroutine 11 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal.go:152 goroutine 816 [IO wait] internal/poll.runtime_pollWait(0x104b75000, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000694a20?, 0x1400095c000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000694a20, {0x1400095c000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000702a70, {0x1400095c000?, 0x14000fb6da8?, 0x1004c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102612d80, 0x14000241180}, {0x102611ec8, 0x140000a2000}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000fb6e01?, {0x102612d80, 0x14000241180}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x103aad230?, {0x102612d80?, 0x14000241180?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102612d80, 0x14000241180}, {0x102611fc0, 0x14000702a70}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140005b4000?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 814 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 66 [select] github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts.func2(0x140006e8000?) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:131 github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts in goroutine 11 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:128 goroutine 82 [select, 123 minutes] github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal.func1() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:32 github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal in goroutine 11 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:30 goroutine 811 [select, 12 minutes] os/exec.(*Cmd).watchCtx(0x140005b4300, 0x1400070cee0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 761 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 800 [IO wait, 12 minutes] internal/poll.runtime_pollWait(0x104b73a00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000caa3c0?, 0x1400094a000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000caa3c0, {0x1400094a000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140005a6040, {0x1400094a000?, 0x14000fba5a8?, 0x1004c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102612d80, 0x14000bc80c0}, {0x102611ec8, 0x140005a6068}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000496a01?, {0x102612d80, 0x14000bc80c0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x103aad230?, {0x102612d80?, 0x14000bc80c0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102612d80, 0x14000bc80c0}, {0x102611fc0, 0x140005a6040}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140008e8008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 762 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 815 [IO wait] internal/poll.runtime_pollWait(0x104b73400, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000694960?, 0x1400092e000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000694960, {0x1400092e000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000702a58, {0x1400092e000?, 0x140009e05a8?, 0x1004c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102612d80, 0x14000241100}, {0x102611ec8, 0x14000124000}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000118301?, {0x102612d80, 0x14000241100}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x103aad230?, {0x102612d80?, 0x14000241100?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102612d80, 0x14000241100}, {0x102611fc0, 0x14000702a58}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140001191f0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 814 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 817 [select] os/exec.(*Cmd).watchCtx(0x140005b4000, 0x14000119340) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 814 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 801 [select, 12 minutes] os/exec.(*Cmd).watchCtx(0x140002cc000, 0x140001180e0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 762 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 809 [IO wait, 12 minutes] internal/poll.runtime_pollWait(0x104b74400, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000694ba0?, 0x14000940000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000694ba0, {0x14000940000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140000a2288, {0x14000940000?, 0x14000fb5da8?, 0x1004c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102612d80, 0x14000241300}, {0x102611ec8, 0x140005a6060}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000fb5e01?, {0x102612d80, 0x14000241300}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x103aad230?, {0x102612d80?, 0x14000241300?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102612d80, 0x14000241300}, {0x102611fc0, 0x140000a2288}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400070c620?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 761 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 799 [IO wait, 12 minutes] internal/poll.runtime_pollWait(0x104b73e00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000caa300?, 0x140009f0000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000caa300, {0x140009f0000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140005a6028, {0x140009f0000?, 0x14000fb55a8?, 0x1004c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102612d80, 0x14000bc8080}, {0x102611ec8, 0x14000702058}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1400070cd01?, {0x102612d80, 0x14000bc8080}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x103aad230?, {0x102612d80?, 0x14000bc8080?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102612d80, 0x14000bc8080}, {0x102611fc0, 0x140005a6028}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140005b4180?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 762 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 810 [IO wait, 12 minutes] internal/poll.runtime_pollWait(0x104b74e00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000694d20?, 0x140009aa000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000694d20, {0x140009aa000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140000a22b8, {0x140009aa000?, 0x14000fb4da8?, 0x1004c7a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102612d80, 0x14000241380}, {0x102611ec8, 0x14000702050}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000496a01?, {0x102612d80, 0x14000241380}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x103aad230?, {0x102612d80?, 0x14000241380?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102612d80, 0x14000241380}, {0x102611fc0, 0x140000a22b8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x14000bfa008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 761 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 Interrupted by User Second interrupt received; Ginkgo will run any reporting nodes but will skip all remaining specs and cleanup nodes. Interrupt again to bail immediately. Here's a current progress report: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 11m47.878s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [AfterEach] (Node Runtime: 2.876s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 Spec Goroutine goroutine 814 [syscall] syscall.syscall6(0x100da7c00?, 0x104bd2228?, 0x1048745c0?, 0x90?, 0x14000680008?, 0x14000a8e6c0?, 0x1400036cb08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400036cb38?, 0x1005866fc?, 0x90?, 0x102582120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029a850?, 0x1400036cb74, 0x14000da7c00?, 0x1400029a7e0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000241200) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000db6c08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140005b4000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140005b4000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400017f9e0?, 0x102619e88?, 0x1400029a690?}}, {0x10263a488?, 0x1400029a690?}, {0x14000aca4d0?, 0xc23316908cec3e28?}, {0x102612d80, 0x14000241100}, {0x102612d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400017f9e0?, 0x103b27728?, 0x103afdfa0?}}, {0x10263a488, 0x1400029a690}, {0x14000aca4d0, 0xb0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.Get({{0x10263ea38?, 0x14000496a38?}, {0x101c57fba?, 0x140009fc9c0?}}, {0x101caf2e2, 0x31}, {{0x0, 0x0, 0x0}, 0x0, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:208 | ctx, cancel := context.WithTimeout(context.Background(), MediumTimeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.SaveTestCaseResources(0x140006cd710, {0x140008997e8, 0x11}, {0x140009ed110, 0x25}, {0x101c54370, 0x4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:765 | errorFileName := fmt.Sprintf("%s/e2e_failed__%s__%s_error.txt", dumpPath, labels["testcase"], additional) | > cmdr := kubectl.Get( | "virtualization,cvi,vmc,intvirt,pod,volumesnapshot", | kc.GetOptions{ > github.com/deckhouse/virtualization/tests/e2e.SaveTestCaseDump(0x140006cd710, {0x101c68e1e, 0x11}, {0x140009ed110, 0x25}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:757 | } | > SaveTestCaseResources(labels, additional, namespace, tmpDir) | SavePodLogsAndDescriptions(labels, additional, namespace, tmpDir) | } > github.com/deckhouse/virtualization/tests/e2e.init.func12.2() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:68 | AfterEach(func() { | if CurrentSpecReport().Failed() { > SaveTestCaseDump(testCaseLabel, CurrentSpecReport().LeafNodeText, ns) | } | }) github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 11 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 761 [syscall, 12 minutes] syscall.syscall6(0x100da6a10?, 0x104bd2228?, 0x1048745c0?, 0x90?, 0x14000580808?, 0x14000a8e360?, 0x14000b33a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000b33a88?, 0x1005866fc?, 0x90?, 0x102582120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029b110?, 0x14000b33ac4, 0x14000da6a10?, 0x1400029b0a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140002413c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000db6408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140005b4300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140005b4300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400017f9e0?, 0x102619e88?, 0x1400029afc0?}}, {0x10263a488?, 0x1400029afc0?}, {0x14000028780?, 0xc23316d81b4235e0?}, {0x102612d80, 0x14000241300}, {0x102612d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400017f9e0?, 0x103b27728?, 0x103afdfa0?}}, {0x10263a488, 0x1400029afc0}, {0x14000028780, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10263ea38?, 0x14000496a38?}, {0x101c57fba?, 0x14000fb4e88?}}, {0x101c96204, 0x28}, {0x140009fc1c0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 805 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 805 [sync.WaitGroup.Wait, 12 minutes] sync.runtime_SemacquireWaitGroup(0x100504b30?, 0xa0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000bfa010) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000be2000, 0x2, 0x101f7de58?}, {0x101c96204, 0x28}, {{0x0, 0x0, 0x0}, 0x140006cd710, {0x140009ed110, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x101c96204, 0x28}, {{0x0, 0x0, 0x0}, 0x140006cd710, {0x140009ed110, 0x25}, {0x140009eea80, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x101c96204, 0x28}, {0x101c54eef, 0x5}, {{0x0, 0x0, 0x0}, 0x140006cd710, {0x140009ed110, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 11 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 goroutine 762 [syscall, 12 minutes] syscall.syscall6(0x100bfa050?, 0x10fe07680?, 0x104874f30?, 0x90?, 0x14000980008?, 0x14000662120?, 0x14000b35a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000b35a88?, 0x1005866fc?, 0x90?, 0x102582120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400027e930?, 0x14000b35ac4, 0x14000bfa050?, 0x1400027e850?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000674300) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140005ca408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140002cc000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140002cc000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x1400017f9e0?, 0x102619e88?, 0x140002c0000?}}, {0x10263a488?, 0x140002c0000?}, {0x140006fe000?, 0xc23316d81b4264c0?}, {0x102612d80, 0x14000bc8080}, {0x102612d80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x1400017f9e0?, 0x103b27728?, 0x103afdfa0?}}, {0x10263a488, 0x140002c0000}, {0x140006fe000, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10263ea38?, 0x14000496a38?}, {0x101c57fba?, 0x140009dd688?}}, {0x101c96204, 0x28}, {0x140009fc1e0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 805 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 20:17:27.094 (2.877s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/artifacts/unique-sds-20251012-153735-20199/junit.xml b/artifacts/unique-sds-20251012-153735-20199/junit.xml new file mode 100644 index 0000000000..79eecbc705 --- /dev/null +++ b/artifacts/unique-sds-20251012-153735-20199/junit.xml @@ -0,0 +1,784 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 15:57:52.653 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 15:57:56.734 (4.081s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 15:57:56.734 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 15:57:56.734 (0s) + + + > Enter [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/12/25 15:57:56.738 < Exit [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/12/25 15:58:00.466 (3.728s) > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/12/25 15:58:00.466 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/12/25 15:58:00.466 (0s) > Enter [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/12/25 15:58:00.466 STEP: Create a vmip automatically and check its binding with a lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:81 @ 10/12/25 15:58:00.466 STEP: Remove label from the lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:87 @ 10/12/25 15:58:00.828 STEP: Wait for the label to be restored by the controller - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:96 @ 10/12/25 15:58:00.941 < Exit [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/12/25 15:58:01.167 (701ms) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/12/25 15:58:01.167 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/12/25 15:58:01.167 (0s) + + + > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/12/25 15:58:01.168 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/12/25 15:58:01.168 (0s) > Enter [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/12/25 15:58:01.168 STEP: Create an intermediate vmip automatically to allocate a new ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:109 @ 10/12/25 15:58:01.168 STEP: Delete the intermediate vmip automatically and check that the lease is released - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:115 @ 10/12/25 15:58:02.153 STEP: Reuse the released lease with a static vmip - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:124 @ 10/12/25 15:58:02.439 STEP: Delete the static vmip and lease, then create another static vmip with this ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:132 @ 10/12/25 15:58:02.807 < Exit [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/12/25 15:58:03.452 (2.284s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/12/25 15:58:03.452 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/12/25 15:58:03.452 (0s) + + + > Enter [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/12/25 15:58:03.452 < Exit [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/12/25 15:58:04.978 (1.526s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/12/25 15:58:04.978 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/12/25 15:58:14.273 (9.295s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/12/25 15:58:14.273 < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/12/25 15:58:14.273 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualimages.virtualization.deckhouse.io head-5073ae15-vi-ubuntu-http --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualimages/head-5073ae15-vi-ubuntu-http\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:89 @ 10/12/25 16:14:56.36 + > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/12/25 15:58:14.273 STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/12/25 15:58:14.273 END STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/12/25 16:14:56.359 (16m42.098s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualimages.virtualization.deckhouse.io head-5073ae15-vi-ubuntu-http --namespace head-5073ae15-end-to-end-affinity-toleration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualimages/head-5073ae15-vi-ubuntu-http\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:89 @ 10/12/25 16:14:56.36 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/12/25 16:14:56.36 (16m42.098s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/12/25 16:14:56.36 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/12/25 16:15:02.505 (6.145s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:120 @ 10/12/25 16:15:02.505 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:283 @ 10/12/25 16:15:02.505 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:372 @ 10/12/25 16:15:02.506 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:467 @ 10/12/25 16:15:02.506 + + + > Enter [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/12/25 16:15:02.506 < Exit [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/12/25 16:15:02.507 (2ms) > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 16:15:02.507 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 16:15:02.507 (0s) > Enter [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/12/25 16:15:02.507 < Exit [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/12/25 16:15:03.996 (1.488s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 16:15:03.996 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 16:15:03.996 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 16:15:03.996 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 16:15:03.996 (0s) > Enter [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/12/25 16:15:03.996 STEP: Project should be deployed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:73 @ 10/12/25 16:15:03.996 < Exit [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/12/25 16:15:06.746 (2.75s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 16:15:06.746 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 16:15:06.746 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 16:15:06.747 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 16:15:06.747 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/12/25 16:15:06.747 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/12/25 16:15:09.829 (3.082s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 16:15:09.829 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 16:15:09.829 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualimages.virtualization.deckhouse.io head-5073ae15-vi --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualimages/head-5073ae15-vi\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/12/25 16:31:51.962 + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 16:15:09.829 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/12/25 16:15:09.829 (0s) > Enter [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/12/25 16:15:09.829 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/12/25 16:15:09.829 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualimages.virtualization.deckhouse.io head-5073ae15-vi --namespace head-5073ae15-end-to-end-importer-network-policy --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualimages/head-5073ae15-vi\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:95 @ 10/12/25 16:31:51.962 < Exit [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/12/25 16:31:51.962 (16m42.131s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 16:31:51.962 The list of pods is empty; nothing to dump. < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/12/25 16:31:57.665 (5.703s) > Enter [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/12/25 16:31:57.665 STEP: Delete manifests - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:50 @ 10/12/25 16:31:57.665 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 16:31:57.665 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 16:32:01.52 (3.854s) < Exit [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/12/25 16:32:01.52 (3.855s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/12/25 16:32:01.52 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/12/25 16:32:01.52 + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1012 16:41:59.532409 80502 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1012 16:42:00.178317 80503 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 @ 10/12/25 16:50:51.352 + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/12/25 16:32:01.52 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/12/25 16:32:03.475 (1.955s) > Enter [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/12/25 16:32:03.475 < Exit [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/12/25 16:32:08.95 (5.475s) > Enter [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/12/25 16:32:08.95 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 @ 10/12/25 16:32:08.95 [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:2, cap:2>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-bios --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1012 16:41:59.532409 80502 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-bios\n\nwaited for: 'jsonpath={.status.phase}=Running'", "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-migration-uefi --namespace head-5073ae15-end-to-end-vm-evacuation --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: W1012 16:42:00.178317 80503 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-migration-uefi\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:98 @ 10/12/25 16:50:51.352 < Exit [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/12/25 16:50:51.352 (16m41.669s) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/12/25 16:50:51.352 The list of pods is empty; nothing to dump. STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 16:50:57.141 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/12/25 16:51:01.934 (4.792s) < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/12/25 16:51:01.934 (10.582s) + + + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/12/25 16:51:01.934 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:52 @ 10/12/25 16:51:03.888 (1.953s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:62 @ 10/12/25 16:51:03.888 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:62 @ 10/12/25 16:51:31.076 (27.189s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 16:51:31.076 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 16:51:31.077 (0s) + + + > Enter [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/12/25 16:51:31.077 < Exit [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:83 @ 10/12/25 16:51:32.152 (1.075s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 16:51:32.152 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 16:51:32.152 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:4, cap:4>: [ "cmd: kubectl wait virtualimages.virtualization.deckhouse.io head-5073ae15-vi-from-vi-ubuntu-http --namespace head-5073ae15-end-to-end-complex-test --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1012 17:14:19.951967 81706 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualimages/head-5073ae15-vi-from-vi-ubuntu-http\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualimages.virtualization.deckhouse.io head-5073ae15-vi-ubuntu-http --namespace head-5073ae15-end-to-end-complex-test --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1012 17:14:19.953166 81704 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualimages/head-5073ae15-vi-ubuntu-http\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualimages.virtualization.deckhouse.io head-5073ae15-vi-alpine-registry --namespace head-5073ae15-end-to-end-complex-test --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1012 17:14:19.940197 81705 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualimages/head-5073ae15-vi-alpine-registry\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualimages.virtualization.deckhouse.io head-5073ae15-vi-from-cvi-ubuntu-http --namespace head-5073ae15-end-to-end-complex-test --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1012 17:14:19.952170 81703 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualimages/head-5073ae15-vi-from-cvi-ubuntu-http\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:103 @ 10/12/25 17:29:44.18 + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/12/25 16:51:32.153 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:102 @ 10/12/25 16:51:32.153 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:4, cap:4>: [ "cmd: kubectl wait virtualimages.virtualization.deckhouse.io head-5073ae15-vi-from-vi-ubuntu-http --namespace head-5073ae15-end-to-end-complex-test --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1012 17:14:19.951967 81706 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualimages/head-5073ae15-vi-from-vi-ubuntu-http\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualimages.virtualization.deckhouse.io head-5073ae15-vi-ubuntu-http --namespace head-5073ae15-end-to-end-complex-test --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1012 17:14:19.953166 81704 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualimages/head-5073ae15-vi-ubuntu-http\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualimages.virtualization.deckhouse.io head-5073ae15-vi-alpine-registry --namespace head-5073ae15-end-to-end-complex-test --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1012 17:14:19.940197 81705 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualimages/head-5073ae15-vi-alpine-registry\n\nwaited for: 'jsonpath={.status.phase}=Ready'", "cmd: kubectl wait virtualimages.virtualization.deckhouse.io head-5073ae15-vi-from-cvi-ubuntu-http --namespace head-5073ae15-end-to-end-complex-test --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1012 17:14:19.952170 81703 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualimages/head-5073ae15-vi-from-cvi-ubuntu-http\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:103 @ 10/12/25 17:29:44.18 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:101 @ 10/12/25 17:29:44.181 (16m42.113s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 17:29:44.181 The list of pods is empty; nothing to dump. < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:46 @ 10/12/25 17:29:50.238 (6.057s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:112 @ 10/12/25 17:29:50.238 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:123 @ 10/12/25 17:29:50.238 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:134 @ 10/12/25 17:29:50.238 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:141 @ 10/12/25 17:29:50.238 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:152 @ 10/12/25 17:29:50.238 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:162 @ 10/12/25 17:29:50.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:173 @ 10/12/25 17:29:50.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:184 @ 10/12/25 17:29:50.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:196 @ 10/12/25 17:29:50.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:220 @ 10/12/25 17:29:50.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:245 @ 10/12/25 17:29:50.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:268 @ 10/12/25 17:29:50.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:280 @ 10/12/25 17:29:50.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:299 @ 10/12/25 17:29:50.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:316 @ 10/12/25 17:29:50.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:341 @ 10/12/25 17:29:50.239 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:354 @ 10/12/25 17:29:50.24 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:358 @ 10/12/25 17:29:50.24 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:375 @ 10/12/25 17:29:50.24 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:388 @ 10/12/25 17:29:50.24 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:405 @ 10/12/25 17:29:50.24 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:444 @ 10/12/25 17:29:50.24 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:479 @ 10/12/25 17:29:50.24 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:497 @ 10/12/25 17:29:50.24 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:512 @ 10/12/25 17:29:50.24 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:528 @ 10/12/25 17:29:50.24 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:545 @ 10/12/25 17:29:50.24 + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.241 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:29:50.241 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.241 (1ms) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.241 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:29:50.242 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.242 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.242 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:29:50.242 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.242 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.242 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:29:50.242 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.242 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.242 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:29:50.242 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.242 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.242 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:29:50.242 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.242 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.243 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:29:50.243 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.243 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.243 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:29:50.243 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.243 (0s) + + + + > Enter [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.243 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/12/25 17:29:50.243 < Exit [BeforeEach] [sig-storage] StorageClassMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:48 @ 10/12/25 17:29:50.243 (0s) + + + > Enter [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/12/25 17:29:50.243 < Exit [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/12/25 17:29:52.19 (1.947s) > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/12/25 17:29:52.19 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/12/25 17:29:52.19 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/12/25 17:29:52.19 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/12/25 17:29:56.254 (4.065s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/12/25 17:29:56.254 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/12/25 17:29:56.254 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-safe --namespace head-5073ae15-end-to-end-vm-restore-safe --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-safe\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:107 @ 10/12/25 17:46:37.96 + > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/12/25 17:29:56.255 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/12/25 17:29:56.255 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/12/25 17:29:56.255 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/12/25 17:29:56.255 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/12/25 17:46:37.96 (16m41.657s) [FAILED] should observe resources in ''jsonpath={.status.phase}=Running'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualmachine.virtualization.deckhouse.io head-5073ae15-vm-restore-safe --namespace head-5073ae15-end-to-end-vm-restore-safe --for='jsonpath={.status.phase}=Running' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualmachines/head-5073ae15-vm-restore-safe\n\nwaited for: 'jsonpath={.status.phase}=Running'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:107 @ 10/12/25 17:46:37.96 < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/12/25 17:46:37.96 (16m41.657s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/12/25 17:46:37.96 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/12/25 17:46:44.207 (6.247s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:115 @ 10/12/25 17:46:44.207 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:158 @ 10/12/25 17:46:44.208 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:348 @ 10/12/25 17:46:44.208 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:374 @ 10/12/25 17:46:44.208 + + + [FAILED] Warning: resource namespaces/head-5073ae15-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "ceph-pool-r2-csi-rbd" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x1400075e340>: exit status 1 { ProcessState: { pid: 84693, status: 256, rusage: { Utime: { Sec: 0, Usec: 294652, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 60432, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 70877184, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5139, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 49, Msgrcv: 54, Nsignals: 338, Nvcsw: 303, Nivcsw: 2716, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/12/25 17:46:51.291 + > Enter [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/12/25 17:46:44.208 < Exit [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/12/25 17:46:47.332 (3.124s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/12/25 17:46:47.332 [FAILED] Warning: resource namespaces/head-5073ae15-end-to-end-image-hotplug is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. Error from server (Forbidden): error when creating "/tmp/testdata/image-hotplug": admission webhook "vi.virtualization-controller.validate.d8-virtualization" denied the request: the storage class "ceph-pool-r2-csi-rbd" lacks of capabilities to support 'Virtual Images on PVC' function; use StorageClass that supports volume mode 'Block' and access mode 'ReadWriteMany' Unexpected error: <*exec.ExitError | 0x1400075e340>: exit status 1 { ProcessState: { pid: 84693, status: 256, rusage: { Utime: { Sec: 0, Usec: 294652, Pad_cgo_0: [0, 0, 0, 0], }, Stime: { Sec: 0, Usec: 60432, Pad_cgo_0: [0, 0, 0, 0], }, Maxrss: 70877184, Ixrss: 0, Idrss: 0, Isrss: 0, Minflt: 5139, Majflt: 3, Nswap: 0, Inblock: 0, Oublock: 0, Msgsnd: 49, Msgrcv: 54, Nsignals: 338, Nvcsw: 303, Nivcsw: 2716, }, }, Stderr: nil, } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:109 @ 10/12/25 17:46:51.291 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/12/25 17:46:51.292 (3.96s) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/12/25 17:46:51.292 The list of pods is empty; nothing to dump. < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/12/25 17:46:56.959 (5.668s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:112 @ 10/12/25 17:46:56.961 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:147 @ 10/12/25 17:46:56.961 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:190 @ 10/12/25 17:46:56.961 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:196 @ 10/12/25 17:46:56.961 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:204 @ 10/12/25 17:46:56.961 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:224 @ 10/12/25 17:46:56.961 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:236 @ 10/12/25 17:46:56.961 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:258 @ 10/12/25 17:46:56.961 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:284 @ 10/12/25 17:46:56.961 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:294 @ 10/12/25 17:46:56.961 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:309 @ 10/12/25 17:46:56.961 + + + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/12/25 17:46:56.962 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/12/25 17:46:58.943 (1.981s) > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 17:46:58.943 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 17:46:58.943 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/12/25 17:46:58.943 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/12/25 17:47:01.962 (3.019s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 17:47:01.962 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 17:47:01.962 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualimages.virtualization.deckhouse.io head-5073ae15-vi-alpine-http --namespace head-5073ae15-end-to-end-vm-label-annotation --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1012 18:10:38.250308 84712 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nW1012 18:43:39.991900 84712 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nW1012 19:46:10.004982 84712 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualimages/head-5073ae15-vi-alpine-http\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:76 @ 10/12/25 19:53:44.073 + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 17:47:01.962 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/12/25 17:47:01.962 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/12/25 17:47:01.962 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:75 @ 10/12/25 17:47:01.962 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualimages.virtualization.deckhouse.io head-5073ae15-vi-alpine-http --namespace head-5073ae15-end-to-end-vm-label-annotation --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: W1012 18:10:38.250308 84712 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nW1012 18:43:39.991900 84712 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nW1012 19:46:10.004982 84712 reflector.go:492] k8s.io/client-go/tools/watch/informerwatcher.go:146: watch of *unstructured.Unstructured ended with: an error on the server (\"unable to decode an event from the watch stream: http2: client connection lost\") has prevented the request from succeeding\nerror: timed out waiting for the condition on virtualimages/head-5073ae15-vi-alpine-http\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:76 @ 10/12/25 19:53:44.073 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/12/25 19:53:44.073 (16m42.119s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 19:53:44.073 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/12/25 19:53:50.346 (6.273s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/12/25 19:53:50.347 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/12/25 19:53:50.347 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/12/25 19:53:50.347 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/12/25 19:53:50.348 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/12/25 19:53:50.348 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/12/25 19:53:50.348 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/12/25 19:53:50.348 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/12/25 19:53:50.348 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/12/25 19:53:50.348 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/12/25 19:53:50.348 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/12/25 19:53:50.348 + + + > Enter [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/12/25 19:53:50.348 < Exit [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/12/25 19:53:52.334 (1.986s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/12/25 19:53:52.334 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/12/25 19:53:57.15 (4.816s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/12/25 19:53:57.15 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/12/25 19:53:57.15 (0s) + + + [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualimages.virtualization.deckhouse.io head-5073ae15-vi-alpine-http --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualimages/head-5073ae15-vi-alpine-http\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:90 @ 10/12/25 20:10:39.292 + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/12/25 19:53:57.15 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:89 @ 10/12/25 19:53:57.15 [FAILED] should observe resources in ''jsonpath={.status.phase}=Ready'' state before 16m40s timeout Expected <[]string | len:1, cap:1>: [ "cmd: kubectl wait virtualimages.virtualization.deckhouse.io head-5073ae15-vi-alpine-http --namespace head-5073ae15-end-to-end-vm-configuration --for='jsonpath={.status.phase}=Ready' --timeout=16m40s\nstderr: error: timed out waiting for the condition on virtualimages/head-5073ae15-vi-alpine-http\n\nwaited for: 'jsonpath={.status.phase}=Ready'", ] to be empty In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:90 @ 10/12/25 20:10:39.292 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/12/25 20:10:39.292 (16m42.141s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/12/25 20:10:39.292 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/12/25 20:10:45.299 (6.007s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/12/25 20:10:45.3 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:109 @ 10/12/25 20:10:45.3 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:123 @ 10/12/25 20:10:45.3 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:147 @ 10/12/25 20:10:45.3 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:161 @ 10/12/25 20:10:45.3 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:183 @ 10/12/25 20:10:45.3 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:202 @ 10/12/25 20:10:45.3 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:226 @ 10/12/25 20:10:45.3 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:240 @ 10/12/25 20:10:45.3 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:250 @ 10/12/25 20:10:45.3 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:265 @ 10/12/25 20:10:45.3 + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/12/25 20:10:47.275 + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/12/25 20:10:45.301 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:71 @ 10/12/25 20:10:47.275 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:55 @ 10/12/25 20:10:47.275 (1.975s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/12/25 20:10:47.275 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:83 @ 10/12/25 20:10:53.691 (6.416s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:100 @ 10/12/25 20:10:53.691 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:111 @ 10/12/25 20:10:53.691 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:122 @ 10/12/25 20:10:53.692 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:133 @ 10/12/25 20:10:53.692 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:144 @ 10/12/25 20:10:53.692 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:164 @ 10/12/25 20:10:53.692 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:185 @ 10/12/25 20:10:53.692 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:229 @ 10/12/25 20:10:53.692 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:290 @ 10/12/25 20:10:53.692 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:341 @ 10/12/25 20:10:53.692 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:367 @ 10/12/25 20:10:53.693 + + + [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:51 @ 10/12/25 20:10:55.673 + > Enter [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/12/25 20:10:53.693 [FAILED] immediate storage class cannot be nil; please set up the immediate storage class in the cluster Expected <*v1.StorageClass | 0x0>: nil not to be nil In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:51 @ 10/12/25 20:10:55.673 < Exit [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:39 @ 10/12/25 20:10:55.673 (1.981s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/12/25 20:10:55.673 The list of pods is empty; nothing to dump. < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:71 @ 10/12/25 20:11:01.297 (5.624s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:94 @ 10/12/25 20:11:01.297 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:103 @ 10/12/25 20:11:01.298 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:114 @ 10/12/25 20:11:01.298 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:128 @ 10/12/25 20:11:01.298 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:139 @ 10/12/25 20:11:01.298 + + + + > Enter [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/12/25 20:11:01.298 [SKIPPED] Module SDN is disabled. Skipping all tests for module SDN. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:50 @ 10/12/25 20:11:01.8 < Exit [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/12/25 20:11:01.801 (503ms) > Enter [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/12/25 20:11:01.801 < Exit [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/12/25 20:11:01.801 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/12/25 20:11:01.802 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/12/25 20:11:01.802 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:111 @ 10/12/25 20:11:01.802 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:125 @ 10/12/25 20:11:01.802 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:141 @ 10/12/25 20:11:01.802 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:156 @ 10/12/25 20:11:01.802 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:169 @ 10/12/25 20:11:01.802 + + + > Enter [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/12/25 20:11:01.803 < Exit [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/12/25 20:11:03.781 (1.978s) > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/12/25 20:11:03.781 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/12/25 20:11:03.781 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/12/25 20:11:03.781 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/12/25 20:11:09.742 (5.961s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/12/25 20:11:09.742 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/12/25 20:11:09.742 (0s) + + + [INTERRUPTED] Interrupted by User In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/12/25 20:17:24.214 This is the Progress Report generated when the interrupt was received: VirtualMachineRestoreForce When the virtualization resources are applied checks the resources phase (Spec Runtime: 6m14.471s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 In [It] (Node Runtime: 6m14.471s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 At [By Step] `VirtualMachine` agent should be ready (Step Runtime: 6m14.471s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 Spec Goroutine goroutine 508 [sync.WaitGroup.Wait, 6 minutes] sync.runtime_SemacquireWaitGroup(0x100ee0b30?, 0x80?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000b871d0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000956de0, 0x2, 0x102959e58?}, {0x102676a2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x102676a2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, 0x29}, {0x140007f1320, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x102676a2d, 0x2a}, {0x102633a0a, 0x7}, {{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, 0x29}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, 0x29}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func22.4.2.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 | It("checks the resources phase", func() { | By("`VirtualMachine` agent should be ready", func() { > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: namespace, github.com/onsi/ginkgo/v2/internal.(*Suite).By(0x140000fd508, {0x10266ddb1, 0x26}, {0x14000b79f48, 0x1, 0x1025c3614?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:323 github.com/onsi/ginkgo/v2.By({0x10266ddb1?, 0x102ff5eb0?}, {0x14000918748?, 0x1400049a410?, 0x14000918758?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:600 > github.com/deckhouse/virtualization/tests/e2e.init.func22.4.2() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 | | It("checks the resources phase", func() { > By("`VirtualMachine` agent should be ready", func() { | WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 512 [syscall, 6 minutes] syscall.syscall6(0x90010100954030?, 0x110a7faa0?, 0x105250f30?, 0x90?, 0x14000580008?, 0x14000968000?, 0x1400071fa58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400071fa88?, 0x100f626fc?, 0x90?, 0x102f5e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000270380?, 0x1400071fac4, 0x14000954030?, 0x140002702a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000434900) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x1400090a408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000908000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000908000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34f0?, 0x102ff5e88?, 0x14000270000?}}, {0x103016488?, 0x14000270000?}, {0x14000e3a0d0?, 0xc233172ab5e082d0?}, {0x102feed80, 0x14000434800}, {0x102feed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34f0?, 0x104503728?, 0x1044d9fa0?}}, {0x103016488, 0x14000270000}, {0x14000e3a0d0, 0xd0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10301aa38?, 0x1400000fde8?}, {0x102633fba?, 0x1400071fe88?}}, {0x102676a2d, 0x2a}, {0x14000b56a00, 0x28}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 508 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 513 [syscall, 6 minutes] syscall.syscall6(0x10067a840?, 0x110a8eaf0?, 0x105250a78?, 0x90?, 0x140000a5008?, 0x14000882090?, 0x1400097ba58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400097ba88?, 0x100f626fc?, 0x90?, 0x102f5e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000490380?, 0x1400097bac4, 0x1400067a840?, 0x140004902a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000596600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140008c4408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140008c0000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140008c0000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34f0?, 0x102ff5e88?, 0x140004900e0?}}, {0x103016488?, 0x140004900e0?}, {0x140008b6000?, 0xc233172ab5e0ec48?}, {0x102feed80, 0x14000596000}, {0x102feed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34f0?, 0x104503728?, 0x1044d9fa0?}}, {0x103016488, 0x140004900e0}, {0x140008b6000, 0xc6}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10301aa38?, 0x1400000fde8?}, {0x102633fba?, 0x14000a20f40?}}, {0x102676a2d, 0x2a}, {0x14000b56a29, 0x1e}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 508 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) Other Goroutines goroutine 5 [running] github.com/onsi/ginkgo/v2/internal.extractRunningGoroutines() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:181 github.com/onsi/ginkgo/v2/internal.NewProgressReport(_, {{0x14000956c00, 0x2, 0x2}, {0x14000255340, 0x2, 0x2}, {0x1400028de30, 0x2, 0x2}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:75 github.com/onsi/ginkgo/v2/internal.(*Suite).generateProgressReport(_, _) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:381 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode(_, {0x17a, 0x4, {0x102655aa3, 0x1a}, 0x140000a2290, {{0x10371dce5, 0x56}, 0x6f, {0x0, ...}, ...}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:980 github.com/onsi/ginkgo/v2/internal.(*group).attemptSpec(0x140007537f0, 0x1, {{0x140006ef508?, 0x14000255340?, 0x2?}, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:199 github.com/onsi/ginkgo/v2/internal.(*group).run(0x140007537f0, {0x14000b58000?, 0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:349 github.com/onsi/ginkgo/v2/internal.(*Suite).runSpecs(0x140000fd508, {0x102630f2b, 0x5}, {0x104503728, 0x0, 0x0}, {0x1400005a0f4, 0x3d}, 0x0, {0x14000092c08, ...}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:489 github.com/onsi/ginkgo/v2/internal.(*Suite).Run(_, {_, _}, {_, _, _}, {_, _}, _, {_, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:130 github.com/onsi/ginkgo/v2.RunSpecs({0x102fee0a0, 0x14000602c40}, {0x102630f2b, 0x5}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:300 github.com/deckhouse/virtualization/tests/e2e.TestE2E(0x14000602c40) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:76 testing.tRunner(0x14000602c40, 0x102fddc50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.(*T).Run in goroutine 1 /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1997 goroutine 1 [chan receive, 126 minutes] testing.(*T).Run(0x14000602a80, {0x10263399a?, 0x140005a1b38?}, 0x102fddc50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2005 testing.runTests.func1(0x14000602a80) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2477 testing.tRunner(0x14000602a80, 0x140005a1c68) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.runTests(0x1400000ed50, {0x104488fd0, 0x1, 0x1}, {0x14000708f20?, 0xf?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2475 testing.(*M).Run(0x140004d6460) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2337 main.main() _testmain.go:45 goroutine 41 [syscall] os/signal.signal_recv() /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sigqueue.go:149 os/signal.loop() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal_unix.go:23 os/signal.Notify.func1.1 in goroutine 5 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal.go:152 goroutine 31 [select, 126 minutes] github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal.func1() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:32 github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:30 goroutine 66 [select] github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts.func2(0x0?) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:131 github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:128 goroutine 396 [IO wait, 6 minutes] internal/poll.runtime_pollWait(0x110980400, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x1400078a2a0?, 0x14000a4c000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x1400078a2a0, {0x14000a4c000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140004280c8, {0x14000a4c000?, 0x14000978da8?, 0x100ea3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102feed80, 0x14000434880}, {0x102fedec8, 0x140008b4210}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x102feed80, 0x14000434880}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x104489230?, {0x102feed80?, 0x14000434880?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102feed80, 0x14000434880}, {0x102fedfc0, 0x140004280c8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 512 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 442 [select, 6 minutes] os/exec.(*Cmd).watchCtx(0x140008c0000, 0x1400098c070) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 513 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 397 [select, 6 minutes] os/exec.(*Cmd).watchCtx(0x14000908000, 0x1400011a620) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 512 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 441 [IO wait, 6 minutes] internal/poll.runtime_pollWait(0x11075d600, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000074240?, 0x14000a34000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000074240, {0x14000a34000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140001261e8, {0x14000a34000?, 0x14000512da8?, 0x100ea3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102feed80, 0x14000596580}, {0x102fedec8, 0x140008b4208}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000512e01?, {0x102feed80, 0x14000596580}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x104489230?, {0x102feed80?, 0x14000596580?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102feed80, 0x14000596580}, {0x102fedfc0, 0x140001261e8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400023e070?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 513 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 395 [IO wait, 6 minutes] internal/poll.runtime_pollWait(0x11075e000, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x1400078a1e0?, 0x14000a3c000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x1400078a1e0, {0x14000a3c000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000428068, {0x14000a3c000?, 0x14000977da8?, 0x100ea3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102feed80, 0x14000434800}, {0x102fedec8, 0x1400049e6e0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x102feed80, 0x14000434800}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x104489230?, {0x102feed80?, 0x14000434800?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102feed80, 0x14000434800}, {0x102fedfc0, 0x14000428068}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 512 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 440 [IO wait, 6 minutes] internal/poll.runtime_pollWait(0x11097fc00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000074060?, 0x14000a22000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000074060, {0x14000a22000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140001260d0, {0x14000a22000?, 0x14000515da8?, 0x100ea3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102feed80, 0x14000596000}, {0x102fedec8, 0x1400049e6d8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000515e01?, {0x102feed80, 0x14000596000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x104489230?, {0x102feed80?, 0x14000596000?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102feed80, 0x14000596000}, {0x102fedfc0, 0x140001260d0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x14000a05be0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 513 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/12/25 20:11:09.742 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/12/25 20:11:09.742 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/12/25 20:11:09.742 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/12/25 20:11:09.742 [INTERRUPTED] Interrupted by User In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/12/25 20:17:24.214 This is the Progress Report generated when the interrupt was received: VirtualMachineRestoreForce When the virtualization resources are applied checks the resources phase (Spec Runtime: 6m14.471s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 In [It] (Node Runtime: 6m14.471s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 At [By Step] `VirtualMachine` agent should be ready (Step Runtime: 6m14.471s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 Spec Goroutine goroutine 508 [sync.WaitGroup.Wait, 6 minutes] sync.runtime_SemacquireWaitGroup(0x100ee0b30?, 0x80?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000b871d0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000956de0, 0x2, 0x102959e58?}, {0x102676a2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x102676a2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, 0x29}, {0x140007f1320, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x102676a2d, 0x2a}, {0x102633a0a, 0x7}, {{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, 0x29}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, 0x29}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func22.4.2.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 | It("checks the resources phase", func() { | By("`VirtualMachine` agent should be ready", func() { > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: namespace, github.com/onsi/ginkgo/v2/internal.(*Suite).By(0x140000fd508, {0x10266ddb1, 0x26}, {0x14000b79f48, 0x1, 0x1025c3614?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:323 github.com/onsi/ginkgo/v2.By({0x10266ddb1?, 0x102ff5eb0?}, {0x14000918748?, 0x1400049a410?, 0x14000918758?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:600 > github.com/deckhouse/virtualization/tests/e2e.init.func22.4.2() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 | | It("checks the resources phase", func() { > By("`VirtualMachine` agent should be ready", func() { | WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 512 [syscall, 6 minutes] syscall.syscall6(0x90010100954030?, 0x110a7faa0?, 0x105250f30?, 0x90?, 0x14000580008?, 0x14000968000?, 0x1400071fa58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400071fa88?, 0x100f626fc?, 0x90?, 0x102f5e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000270380?, 0x1400071fac4, 0x14000954030?, 0x140002702a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000434900) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x1400090a408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000908000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000908000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34f0?, 0x102ff5e88?, 0x14000270000?}}, {0x103016488?, 0x14000270000?}, {0x14000e3a0d0?, 0xc233172ab5e082d0?}, {0x102feed80, 0x14000434800}, {0x102feed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34f0?, 0x104503728?, 0x1044d9fa0?}}, {0x103016488, 0x14000270000}, {0x14000e3a0d0, 0xd0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10301aa38?, 0x1400000fde8?}, {0x102633fba?, 0x1400071fe88?}}, {0x102676a2d, 0x2a}, {0x14000b56a00, 0x28}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 508 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 513 [syscall, 6 minutes] syscall.syscall6(0x10067a840?, 0x110a8eaf0?, 0x105250a78?, 0x90?, 0x140000a5008?, 0x14000882090?, 0x1400097ba58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400097ba88?, 0x100f626fc?, 0x90?, 0x102f5e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000490380?, 0x1400097bac4, 0x1400067a840?, 0x140004902a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000596600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140008c4408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140008c0000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140008c0000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34f0?, 0x102ff5e88?, 0x140004900e0?}}, {0x103016488?, 0x140004900e0?}, {0x140008b6000?, 0xc233172ab5e0ec48?}, {0x102feed80, 0x14000596000}, {0x102feed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34f0?, 0x104503728?, 0x1044d9fa0?}}, {0x103016488, 0x140004900e0}, {0x140008b6000, 0xc6}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10301aa38?, 0x1400000fde8?}, {0x102633fba?, 0x14000a20f40?}}, {0x102676a2d, 0x2a}, {0x14000b56a29, 0x1e}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 508 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) Other Goroutines goroutine 5 [running] github.com/onsi/ginkgo/v2/internal.extractRunningGoroutines() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:181 github.com/onsi/ginkgo/v2/internal.NewProgressReport(_, {{0x14000956c00, 0x2, 0x2}, {0x14000255340, 0x2, 0x2}, {0x1400028de30, 0x2, 0x2}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:75 github.com/onsi/ginkgo/v2/internal.(*Suite).generateProgressReport(_, _) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:381 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode(_, {0x17a, 0x4, {0x102655aa3, 0x1a}, 0x140000a2290, {{0x10371dce5, 0x56}, 0x6f, {0x0, ...}, ...}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:980 github.com/onsi/ginkgo/v2/internal.(*group).attemptSpec(0x140007537f0, 0x1, {{0x140006ef508?, 0x14000255340?, 0x2?}, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:199 github.com/onsi/ginkgo/v2/internal.(*group).run(0x140007537f0, {0x14000b58000?, 0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:349 github.com/onsi/ginkgo/v2/internal.(*Suite).runSpecs(0x140000fd508, {0x102630f2b, 0x5}, {0x104503728, 0x0, 0x0}, {0x1400005a0f4, 0x3d}, 0x0, {0x14000092c08, ...}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:489 github.com/onsi/ginkgo/v2/internal.(*Suite).Run(_, {_, _}, {_, _, _}, {_, _}, _, {_, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:130 github.com/onsi/ginkgo/v2.RunSpecs({0x102fee0a0, 0x14000602c40}, {0x102630f2b, 0x5}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:300 github.com/deckhouse/virtualization/tests/e2e.TestE2E(0x14000602c40) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:76 testing.tRunner(0x14000602c40, 0x102fddc50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.(*T).Run in goroutine 1 /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1997 goroutine 1 [chan receive, 126 minutes] testing.(*T).Run(0x14000602a80, {0x10263399a?, 0x140005a1b38?}, 0x102fddc50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2005 testing.runTests.func1(0x14000602a80) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2477 testing.tRunner(0x14000602a80, 0x140005a1c68) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.runTests(0x1400000ed50, {0x104488fd0, 0x1, 0x1}, {0x14000708f20?, 0xf?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2475 testing.(*M).Run(0x140004d6460) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2337 main.main() _testmain.go:45 goroutine 41 [syscall] os/signal.signal_recv() /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sigqueue.go:149 os/signal.loop() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal_unix.go:23 os/signal.Notify.func1.1 in goroutine 5 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal.go:152 goroutine 31 [select, 126 minutes] github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal.func1() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:32 github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:30 goroutine 66 [select] github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts.func2(0x0?) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:131 github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:128 goroutine 396 [IO wait, 6 minutes] internal/poll.runtime_pollWait(0x110980400, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x1400078a2a0?, 0x14000a4c000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x1400078a2a0, {0x14000a4c000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140004280c8, {0x14000a4c000?, 0x14000978da8?, 0x100ea3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102feed80, 0x14000434880}, {0x102fedec8, 0x140008b4210}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x102feed80, 0x14000434880}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x104489230?, {0x102feed80?, 0x14000434880?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102feed80, 0x14000434880}, {0x102fedfc0, 0x140004280c8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 512 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 442 [select, 6 minutes] os/exec.(*Cmd).watchCtx(0x140008c0000, 0x1400098c070) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 513 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 397 [select, 6 minutes] os/exec.(*Cmd).watchCtx(0x14000908000, 0x1400011a620) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 512 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 441 [IO wait, 6 minutes] internal/poll.runtime_pollWait(0x11075d600, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000074240?, 0x14000a34000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000074240, {0x14000a34000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140001261e8, {0x14000a34000?, 0x14000512da8?, 0x100ea3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102feed80, 0x14000596580}, {0x102fedec8, 0x140008b4208}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000512e01?, {0x102feed80, 0x14000596580}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x104489230?, {0x102feed80?, 0x14000596580?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102feed80, 0x14000596580}, {0x102fedfc0, 0x140001261e8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400023e070?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 513 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 395 [IO wait, 6 minutes] internal/poll.runtime_pollWait(0x11075e000, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x1400078a1e0?, 0x14000a3c000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x1400078a1e0, {0x14000a3c000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000428068, {0x14000a3c000?, 0x14000977da8?, 0x100ea3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102feed80, 0x14000434800}, {0x102fedec8, 0x1400049e6e0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x102feed80, 0x14000434800}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x104489230?, {0x102feed80?, 0x14000434800?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102feed80, 0x14000434800}, {0x102fedfc0, 0x14000428068}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 512 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 440 [IO wait, 6 minutes] internal/poll.runtime_pollWait(0x11097fc00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000074060?, 0x14000a22000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000074060, {0x14000a22000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140001260d0, {0x14000a22000?, 0x14000515da8?, 0x100ea3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102feed80, 0x14000596000}, {0x102fedec8, 0x1400049e6d8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000515e01?, {0x102feed80, 0x14000596000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x104489230?, {0x102feed80?, 0x14000596000?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102feed80, 0x14000596000}, {0x102fedfc0, 0x140001260d0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x14000a05be0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 513 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 Interrupted by User First interrupt received; Ginkgo will run any cleanup and reporting nodes but will skip all remaining specs. Interrupt again to skip cleanup. Here's a current progress report: VirtualMachineRestoreForce When the virtualization resources are applied checks the resources phase (Spec Runtime: 6m14.471s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 In [It] (Node Runtime: 6m14.471s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 At [By Step] `VirtualMachine` agent should be ready (Step Runtime: 6m14.471s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 Spec Goroutine goroutine 508 [sync.WaitGroup.Wait, 6 minutes] sync.runtime_SemacquireWaitGroup(0x100ee0b30?, 0x80?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000b871d0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000956de0, 0x2, 0x102959e58?}, {0x102676a2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x102676a2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, 0x29}, {0x140007f1320, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x102676a2d, 0x2a}, {0x102633a0a, 0x7}, {{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, 0x29}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, 0x29}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func22.4.2.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 | It("checks the resources phase", func() { | By("`VirtualMachine` agent should be ready", func() { > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: namespace, github.com/onsi/ginkgo/v2/internal.(*Suite).By(0x140000fd508, {0x10266ddb1, 0x26}, {0x14000b79f48, 0x1, 0x1025c3614?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:323 github.com/onsi/ginkgo/v2.By({0x10266ddb1?, 0x102ff5eb0?}, {0x14000918748?, 0x1400049a410?, 0x14000918758?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:600 > github.com/deckhouse/virtualization/tests/e2e.init.func22.4.2() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 | | It("checks the resources phase", func() { > By("`VirtualMachine` agent should be ready", func() { | WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 512 [syscall, 6 minutes] syscall.syscall6(0x90010100954030?, 0x110a7faa0?, 0x105250f30?, 0x90?, 0x14000580008?, 0x14000968000?, 0x1400071fa58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400071fa88?, 0x100f626fc?, 0x90?, 0x102f5e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000270380?, 0x1400071fac4, 0x14000954030?, 0x140002702a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000434900) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x1400090a408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000908000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000908000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34f0?, 0x102ff5e88?, 0x14000270000?}}, {0x103016488?, 0x14000270000?}, {0x14000e3a0d0?, 0xc233172ab5e082d0?}, {0x102feed80, 0x14000434800}, {0x102feed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34f0?, 0x104503728?, 0x1044d9fa0?}}, {0x103016488, 0x14000270000}, {0x14000e3a0d0, 0xd0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10301aa38?, 0x1400000fde8?}, {0x102633fba?, 0x1400071fe88?}}, {0x102676a2d, 0x2a}, {0x14000b56a00, 0x28}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 508 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 513 [syscall, 6 minutes] syscall.syscall6(0x10067a840?, 0x110a8eaf0?, 0x105250a78?, 0x90?, 0x140000a5008?, 0x14000882090?, 0x1400097ba58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400097ba88?, 0x100f626fc?, 0x90?, 0x102f5e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000490380?, 0x1400097bac4, 0x1400067a840?, 0x140004902a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000596600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140008c4408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140008c0000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140008c0000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34f0?, 0x102ff5e88?, 0x140004900e0?}}, {0x103016488?, 0x140004900e0?}, {0x140008b6000?, 0xc233172ab5e0ec48?}, {0x102feed80, 0x14000596000}, {0x102feed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34f0?, 0x104503728?, 0x1044d9fa0?}}, {0x103016488, 0x140004900e0}, {0x140008b6000, 0xc6}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10301aa38?, 0x1400000fde8?}, {0x102633fba?, 0x14000a20f40?}}, {0x102676a2d, 0x2a}, {0x14000b56a29, 0x1e}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 508 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/12/25 20:17:24.216 (6m14.473s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/12/25 20:17:24.216 [INTERRUPTED] Interrupted by User In [AfterEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/12/25 20:17:27.092 This is the Progress Report generated when the interrupt was received: VirtualMachineRestoreForce When the virtualization resources are applied checks the resources phase (Spec Runtime: 6m17.35s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 In [AfterEach] (Node Runtime: 2.876s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 Spec Goroutine goroutine 456 [syscall] syscall.syscall6(0x10090dc70?, 0x110a8eaf0?, 0x1052505c0?, 0x90?, 0x14000500808?, 0x14000882b40?, 0x1400002eb08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400002eb38?, 0x100f626fc?, 0x90?, 0x102f5e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000490b60?, 0x1400002eb74, 0x1400090dc70?, 0x14000490a10?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000597200) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140008c4c08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140008c0180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140008c0180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34f0?, 0x102ff5e88?, 0x14000490850?}}, {0x103016488?, 0x14000490850?}, {0x14000884540?, 0xc23316908cec3270?}, {0x102feed80, 0x14000597080}, {0x102feed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34f0?, 0x104503728?, 0x1044d9fa0?}}, {0x103016488, 0x14000490850}, {0x14000884540, 0xb5}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.Get({{0x10301aa38?, 0x1400000fde8?}, {0x102633fba?, 0x140004373b0?}}, {0x10268b2e2, 0x31}, {{0x0, 0x0, 0x0}, 0x0, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:208 | ctx, cancel := context.WithTimeout(context.Background(), MediumTimeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.SaveTestCaseResources(0x1400070b350, {0x14000a15c40, 0x1a}, {0x14000bebd70, 0x29}, {0x102630370, 0x4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:765 | errorFileName := fmt.Sprintf("%s/e2e_failed__%s__%s_error.txt", dumpPath, labels["testcase"], additional) | > cmdr := kubectl.Get( | "virtualization,cvi,vmc,intvirt,pod,volumesnapshot", | kc.GetOptions{ > github.com/deckhouse/virtualization/tests/e2e.SaveTestCaseDump(0x1400070b350, {0x102655aa3, 0x1a}, {0x14000bebd70, 0x29}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:757 | } | > SaveTestCaseResources(labels, additional, namespace, tmpDir) | SavePodLogsAndDescriptions(labels, additional, namespace, tmpDir) | } > github.com/deckhouse/virtualization/tests/e2e.init.func22.3() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:73 | AfterEach(func() { | if CurrentSpecReport().Failed() { > SaveTestCaseDump(testCaseLabel, CurrentSpecReport().LeafNodeText, namespace) | } | github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 508 [sync.WaitGroup.Wait, 6 minutes] sync.runtime_SemacquireWaitGroup(0x100ee0b30?, 0x80?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000b871d0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000956de0, 0x2, 0x102959e58?}, {0x102676a2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x102676a2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, 0x29}, {0x140007f1320, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x102676a2d, 0x2a}, {0x102633a0a, 0x7}, {{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, 0x29}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, 0x29}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func22.4.2.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 | It("checks the resources phase", func() { | By("`VirtualMachine` agent should be ready", func() { > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: namespace, github.com/onsi/ginkgo/v2/internal.(*Suite).By(0x140000fd508, {0x10266ddb1, 0x26}, {0x14000b79f48, 0x1, 0x1025c3614?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:323 github.com/onsi/ginkgo/v2.By({0x10266ddb1?, 0x102ff5eb0?}, {0x14000918748?, 0x1400049a410?, 0x14000918758?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:600 > github.com/deckhouse/virtualization/tests/e2e.init.func22.4.2() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 | | It("checks the resources phase", func() { > By("`VirtualMachine` agent should be ready", func() { | WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 goroutine 512 [syscall, 6 minutes] syscall.syscall6(0x90010100954030?, 0x110a7faa0?, 0x105250f30?, 0x90?, 0x14000580008?, 0x14000968000?, 0x1400071fa58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400071fa88?, 0x100f626fc?, 0x90?, 0x102f5e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000270380?, 0x1400071fac4, 0x14000954030?, 0x140002702a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000434900) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x1400090a408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000908000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000908000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34f0?, 0x102ff5e88?, 0x14000270000?}}, {0x103016488?, 0x14000270000?}, {0x14000e3a0d0?, 0xc233172ab5e082d0?}, {0x102feed80, 0x14000434800}, {0x102feed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34f0?, 0x104503728?, 0x1044d9fa0?}}, {0x103016488, 0x14000270000}, {0x14000e3a0d0, 0xd0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10301aa38?, 0x1400000fde8?}, {0x102633fba?, 0x1400071fe88?}}, {0x102676a2d, 0x2a}, {0x14000b56a00, 0x28}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 508 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 513 [syscall, 6 minutes] syscall.syscall6(0x10067a840?, 0x110a8eaf0?, 0x105250a78?, 0x90?, 0x140000a5008?, 0x14000882090?, 0x1400097ba58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400097ba88?, 0x100f626fc?, 0x90?, 0x102f5e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000490380?, 0x1400097bac4, 0x1400067a840?, 0x140004902a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000596600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140008c4408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140008c0000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140008c0000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34f0?, 0x102ff5e88?, 0x140004900e0?}}, {0x103016488?, 0x140004900e0?}, {0x140008b6000?, 0xc233172ab5e0ec48?}, {0x102feed80, 0x14000596000}, {0x102feed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34f0?, 0x104503728?, 0x1044d9fa0?}}, {0x103016488, 0x140004900e0}, {0x140008b6000, 0xc6}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10301aa38?, 0x1400000fde8?}, {0x102633fba?, 0x14000a20f40?}}, {0x102676a2d, 0x2a}, {0x14000b56a29, 0x1e}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 508 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) Other Goroutines goroutine 5 [running] github.com/onsi/ginkgo/v2/internal.extractRunningGoroutines() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:181 github.com/onsi/ginkgo/v2/internal.NewProgressReport(_, {{0x14000956c00, 0x2, 0x2}, {0x14000255340, 0x2, 0x2}, {0x1400028de30, 0x2, 0x2}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:75 github.com/onsi/ginkgo/v2/internal.(*Suite).generateProgressReport(_, _) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:381 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode(_, {0x177, 0x20, {0x0, 0x0}, 0x140000a21c0, {{0x10371dce5, 0x56}, 0x47, {0x0, ...}, ...}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:980 github.com/onsi/ginkgo/v2/internal.(*group).attemptSpec(0x140007537f0, 0x1, {{0x140006ef508?, 0x14000255340?, 0x2?}, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:286 github.com/onsi/ginkgo/v2/internal.(*group).run(0x140007537f0, {0x14000b58000?, 0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:349 github.com/onsi/ginkgo/v2/internal.(*Suite).runSpecs(0x140000fd508, {0x102630f2b, 0x5}, {0x104503728, 0x0, 0x0}, {0x1400005a0f4, 0x3d}, 0x0, {0x14000092c08, ...}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:489 github.com/onsi/ginkgo/v2/internal.(*Suite).Run(_, {_, _}, {_, _, _}, {_, _}, _, {_, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:130 github.com/onsi/ginkgo/v2.RunSpecs({0x102fee0a0, 0x14000602c40}, {0x102630f2b, 0x5}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:300 github.com/deckhouse/virtualization/tests/e2e.TestE2E(0x14000602c40) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:76 testing.tRunner(0x14000602c40, 0x102fddc50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.(*T).Run in goroutine 1 /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1997 goroutine 1 [chan receive, 126 minutes] testing.(*T).Run(0x14000602a80, {0x10263399a?, 0x140005a1b38?}, 0x102fddc50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2005 testing.runTests.func1(0x14000602a80) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2477 testing.tRunner(0x14000602a80, 0x140005a1c68) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.runTests(0x1400000ed50, {0x104488fd0, 0x1, 0x1}, {0x14000708f20?, 0xf?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2475 testing.(*M).Run(0x140004d6460) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2337 main.main() _testmain.go:45 goroutine 41 [syscall] os/signal.signal_recv() /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sigqueue.go:149 os/signal.loop() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal_unix.go:23 os/signal.Notify.func1.1 in goroutine 5 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal.go:152 goroutine 31 [select, 126 minutes] github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal.func1() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:32 github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:30 goroutine 66 [select] github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts.func2(0x0?) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:131 github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:128 goroutine 459 [select] os/exec.(*Cmd).watchCtx(0x140008c0180, 0x14000647490) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 456 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 458 [IO wait] internal/poll.runtime_pollWait(0x110980000, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140000753e0?, 0x1400083c000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140000753e0, {0x1400083c000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140001262a0, {0x1400083c000?, 0x1400091a5a8?, 0x100ea3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102feed80, 0x14000597140}, {0x102fedec8, 0x1400049e000}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x140000f0101?, {0x102feed80, 0x14000597140}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x104489230?, {0x102feed80?, 0x14000597140?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102feed80, 0x14000597140}, {0x102fedfc0, 0x140001262a0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140004ca000?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 456 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 396 [IO wait, 6 minutes] internal/poll.runtime_pollWait(0x110980400, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x1400078a2a0?, 0x14000a4c000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x1400078a2a0, {0x14000a4c000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140004280c8, {0x14000a4c000?, 0x14000978da8?, 0x100ea3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102feed80, 0x14000434880}, {0x102fedec8, 0x140008b4210}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x102feed80, 0x14000434880}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x104489230?, {0x102feed80?, 0x14000434880?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102feed80, 0x14000434880}, {0x102fedfc0, 0x140004280c8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 512 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 442 [select, 6 minutes] os/exec.(*Cmd).watchCtx(0x140008c0000, 0x1400098c070) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 513 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 397 [select, 6 minutes] os/exec.(*Cmd).watchCtx(0x14000908000, 0x1400011a620) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 512 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 457 [IO wait] internal/poll.runtime_pollWait(0x110980200, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000074cc0?, 0x140007a6000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000074cc0, {0x140007a6000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000126218, {0x140007a6000?, 0x1400091b5a8?, 0x100ea3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102feed80, 0x14000597080}, {0x102fedec8, 0x14000592020}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x140000f0901?, {0x102feed80, 0x14000597080}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x104489230?, {0x102feed80?, 0x14000597080?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102feed80, 0x14000597080}, {0x102fedfc0, 0x14000126218}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140006473b0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 456 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 441 [IO wait, 6 minutes] internal/poll.runtime_pollWait(0x11075d600, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000074240?, 0x14000a34000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000074240, {0x14000a34000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140001261e8, {0x14000a34000?, 0x14000512da8?, 0x100ea3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102feed80, 0x14000596580}, {0x102fedec8, 0x140008b4208}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000512e01?, {0x102feed80, 0x14000596580}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x104489230?, {0x102feed80?, 0x14000596580?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102feed80, 0x14000596580}, {0x102fedfc0, 0x140001261e8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400023e070?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 513 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 395 [IO wait, 6 minutes] internal/poll.runtime_pollWait(0x11075e000, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x1400078a1e0?, 0x14000a3c000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x1400078a1e0, {0x14000a3c000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000428068, {0x14000a3c000?, 0x14000977da8?, 0x100ea3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102feed80, 0x14000434800}, {0x102fedec8, 0x1400049e6e0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x102feed80, 0x14000434800}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x104489230?, {0x102feed80?, 0x14000434800?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102feed80, 0x14000434800}, {0x102fedfc0, 0x14000428068}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 512 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 440 [IO wait, 6 minutes] internal/poll.runtime_pollWait(0x11097fc00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000074060?, 0x14000a22000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000074060, {0x14000a22000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140001260d0, {0x14000a22000?, 0x14000515da8?, 0x100ea3a18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x102feed80, 0x14000596000}, {0x102fedec8, 0x1400049e6d8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x14000515e01?, {0x102feed80, 0x14000596000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x104489230?, {0x102feed80?, 0x14000596000?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x102feed80, 0x14000596000}, {0x102fedfc0, 0x140001260d0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x14000a05be0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 513 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 Interrupted by User Second interrupt received; Ginkgo will run any reporting nodes but will skip all remaining specs and cleanup nodes. Interrupt again to bail immediately. Here's a current progress report: VirtualMachineRestoreForce When the virtualization resources are applied checks the resources phase (Spec Runtime: 6m17.35s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 In [AfterEach] (Node Runtime: 2.876s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 Spec Goroutine goroutine 456 [syscall] syscall.syscall6(0x10090dc70?, 0x110a8eaf0?, 0x1052505c0?, 0x90?, 0x14000500808?, 0x14000882b40?, 0x1400002eb08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400002eb38?, 0x100f626fc?, 0x90?, 0x102f5e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000490b60?, 0x1400002eb74, 0x1400090dc70?, 0x14000490a10?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000597200) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140008c4c08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140008c0180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140008c0180) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34f0?, 0x102ff5e88?, 0x14000490850?}}, {0x103016488?, 0x14000490850?}, {0x14000884540?, 0xc23316908cec3270?}, {0x102feed80, 0x14000597080}, {0x102feed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34f0?, 0x104503728?, 0x1044d9fa0?}}, {0x103016488, 0x14000490850}, {0x14000884540, 0xb5}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.Get({{0x10301aa38?, 0x1400000fde8?}, {0x102633fba?, 0x140004373b0?}}, {0x10268b2e2, 0x31}, {{0x0, 0x0, 0x0}, 0x0, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:208 | ctx, cancel := context.WithTimeout(context.Background(), MediumTimeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.SaveTestCaseResources(0x1400070b350, {0x14000a15c40, 0x1a}, {0x14000bebd70, 0x29}, {0x102630370, 0x4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:765 | errorFileName := fmt.Sprintf("%s/e2e_failed__%s__%s_error.txt", dumpPath, labels["testcase"], additional) | > cmdr := kubectl.Get( | "virtualization,cvi,vmc,intvirt,pod,volumesnapshot", | kc.GetOptions{ > github.com/deckhouse/virtualization/tests/e2e.SaveTestCaseDump(0x1400070b350, {0x102655aa3, 0x1a}, {0x14000bebd70, 0x29}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:757 | } | > SaveTestCaseResources(labels, additional, namespace, tmpDir) | SavePodLogsAndDescriptions(labels, additional, namespace, tmpDir) | } > github.com/deckhouse/virtualization/tests/e2e.init.func22.3() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:73 | AfterEach(func() { | if CurrentSpecReport().Failed() { > SaveTestCaseDump(testCaseLabel, CurrentSpecReport().LeafNodeText, namespace) | } | github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 508 [sync.WaitGroup.Wait, 6 minutes] sync.runtime_SemacquireWaitGroup(0x100ee0b30?, 0x80?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x14000b871d0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000956de0, 0x2, 0x102959e58?}, {0x102676a2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x102676a2d, 0x2a}, {{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, 0x29}, {0x140007f1320, 0x22}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x102676a2d, 0x2a}, {0x102633a0a, 0x7}, {{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, 0x29}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitVMAgentReady({{0x0, 0x0, 0x0}, 0x1400070b350, {0x14000bebd70, 0x29}, {0x0, 0x0}, 0xe8d4a51000}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:274 | func WaitVMAgentReady(opts kc.WaitOptions) { | GinkgoHelper() > WaitPhaseByLabel(kc.ResourceVM, PhaseRunning, opts) | WaitConditionIsTrueByLabel(kc.ResourceVM, vmcondition.TypeAgentReady.String(), opts) | } > github.com/deckhouse/virtualization/tests/e2e.init.func22.4.2.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:113 | It("checks the resources phase", func() { | By("`VirtualMachine` agent should be ready", func() { > WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: namespace, github.com/onsi/ginkgo/v2/internal.(*Suite).By(0x140000fd508, {0x10266ddb1, 0x26}, {0x14000b79f48, 0x1, 0x1025c3614?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:323 github.com/onsi/ginkgo/v2.By({0x10266ddb1?, 0x102ff5eb0?}, {0x14000918748?, 0x1400049a410?, 0x14000918758?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:600 > github.com/deckhouse/virtualization/tests/e2e.init.func22.4.2() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 | | It("checks the resources phase", func() { > By("`VirtualMachine` agent should be ready", func() { | WaitVMAgentReady(kc.WaitOptions{ | Labels: testCaseLabel, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 5 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 goroutine 512 [syscall, 6 minutes] syscall.syscall6(0x90010100954030?, 0x110a7faa0?, 0x105250f30?, 0x90?, 0x14000580008?, 0x14000968000?, 0x1400071fa58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400071fa88?, 0x100f626fc?, 0x90?, 0x102f5e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000270380?, 0x1400071fac4, 0x14000954030?, 0x140002702a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000434900) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x1400090a408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000908000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000908000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34f0?, 0x102ff5e88?, 0x14000270000?}}, {0x103016488?, 0x14000270000?}, {0x14000e3a0d0?, 0xc233172ab5e082d0?}, {0x102feed80, 0x14000434800}, {0x102feed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34f0?, 0x104503728?, 0x1044d9fa0?}}, {0x103016488, 0x14000270000}, {0x14000e3a0d0, 0xd0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10301aa38?, 0x1400000fde8?}, {0x102633fba?, 0x1400071fe88?}}, {0x102676a2d, 0x2a}, {0x14000b56a00, 0x28}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 508 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 513 [syscall, 6 minutes] syscall.syscall6(0x10067a840?, 0x110a8eaf0?, 0x105250a78?, 0x90?, 0x140000a5008?, 0x14000882090?, 0x1400097ba58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x1400097ba88?, 0x100f626fc?, 0x90?, 0x102f5e120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000490380?, 0x1400097bac4, 0x1400067a840?, 0x140004902a0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000596600) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x140008c4408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140008c0000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140008c0000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34f0?, 0x102ff5e88?, 0x140004900e0?}}, {0x103016488?, 0x140004900e0?}, {0x140008b6000?, 0xc233172ab5e0ec48?}, {0x102feed80, 0x14000596000}, {0x102feed80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34f0?, 0x104503728?, 0x1044d9fa0?}}, {0x103016488, 0x140004900e0}, {0x140008b6000, 0xc6}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x10301aa38?, 0x1400000fde8?}, {0x102633fba?, 0x14000a20f40?}}, {0x102676a2d, 0x2a}, {0x14000b56a29, 0x1e}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 508 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/12/25 20:17:27.094 (2.877s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/artifacts/updated-sds-20251012-202031-30115/junit.xml b/artifacts/updated-sds-20251012-202031-30115/junit.xml new file mode 100644 index 0000000000..344f3174a7 --- /dev/null +++ b/artifacts/updated-sds-20251012-202031-30115/junit.xml @@ -0,0 +1,671 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 21:50:34.354 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 21:50:39.216 (4.862s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 21:50:39.216 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/12/25 21:50:39.216 (0s) + + + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/12/25 21:50:39.221 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/12/25 21:50:41.243 (2.023s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/12/25 21:50:41.243 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/12/25 21:50:45.767 (4.524s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 21:50:45.767 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 21:50:45.767 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/12/25 21:50:45.768 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:97 @ 10/12/25 21:50:45.768 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/12/25 21:51:32.187 (46.42s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 21:51:32.187 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 21:51:32.188 (0s) + + + [INTERRUPTED] Interrupted by User In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/12/25 21:59:01.568 This is the Progress Report generated when the interrupt was received: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 7m29.366s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [It] (Node Runtime: 7m29.366s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 At [By Step] VDs should be in Ready phase (Step Runtime: 7m29.366s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 Spec Goroutine goroutine 103 [sync.WaitGroup.Wait, 8 minutes] sync.runtime_SemacquireWaitGroup(0x10022cb30?, 0xd0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x1400026f000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000292040, 0x2, 0x101ca5e58?}, {0x1019be204, 0x28}, {{0x0, 0x0, 0x0}, 0x14000796b70, {0x140006d9f50, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1019be204, 0x28}, {{0x0, 0x0, 0x0}, 0x14000796b70, {0x140006d9f50, 0x25}, {0x140006c9a40, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1019be204, 0x28}, {0x10197ceef, 0x5}, {{0x0, 0x0, 0x0}, 0x14000796b70, {0x140006d9f50, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 25 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 149 [syscall, 8 minutes] syscall.syscall6(0x100e60620?, 0x12f9e7060?, 0x10459cf30?, 0x90?, 0x14000273808?, 0x14000208510?, 0x140009baa58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140009baa88?, 0x1002ae6fc?, 0x90?, 0x1022aa120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029c770?, 0x140009baac4, 0x14000e60620?, 0x1400029c700?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000720240) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000e47408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000171680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000171680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x102341e88?, 0x1400029c620?}}, {0x102362488?, 0x1400029c620?}, {0x1400073a540?, 0xc2331d0c5aaf8ee8?}, {0x10233ad80, 0x140007201c0}, {0x10233ad80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x10384f728?, 0x103825fa0?}}, {0x102362488, 0x1400029c620}, {0x1400073a540, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102366a38?, 0x1400000fc68?}, {0x10197ffba?, 0x140004af698?}}, {0x1019be204, 0x28}, {0x140001360c0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 103 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 147 [sync.Cond.Wait, 8 minutes] sync.runtime_notifyListWait(0x140005f4c48, 0x8b) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140005f4c38) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140005f4c30, {0x14000879000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400076d440?}, {0x14000879000?, 0x101e15b40?, 0x14000e63340?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140005aa840) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140005aa840) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000613908) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000613908) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000613908, {0x14000cba963, 0xe569d, 0x100371c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x14000176848, {0x14000cba963, 0xe569d, 0xe569d}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000cba963?, 0x14000e669b8?, 0x14000cba814?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x140009bfda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x1400059a210, {0x12fdcba60, 0x140006ba420}, {0x10233b5a0, 0x14000075320}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 12 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 49 [sync.Cond.Wait, 8 minutes] sync.runtime_notifyListWait(0x14000171cc8, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x14000171cb8) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x14000171cb0, {0x14000878000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400076d440?}, {0x14000878000?, 0x101e15b40?, 0x14000e188f0?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140005aa780) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140005aa780) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000612608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000612608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000612608, {0x14000ba3174, 0xfce8c, 0x100371c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x140001762c8, {0x14000ba3174, 0xfce8c, 0xfce8c}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000ba3174?, 0x14000e1a778?, 0x14000ba30f2?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x14000901da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x1400059a210, {0x12fdcba60, 0x1400076d230}, {0x10233b5a0, 0x14000075320}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 12 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 150 [syscall, 8 minutes] syscall.syscall6(0x10026f040?, 0x12fbbea38?, 0x10459ca78?, 0x90?, 0x14000790008?, 0x14000f20090?, 0x14000606a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000606a88?, 0x1002ae6fc?, 0x90?, 0x1022aa120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140004821c0?, 0x14000606ac4, 0x1400026f040?, 0x14000482150?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000470c40) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000f28408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000f24000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000f24000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x102341e88?, 0x14000482000?}}, {0x102362488?, 0x14000482000?}, {0x14000f22000?, 0xc2331d0c5aafa658?}, {0x10233ad80, 0x14000470a40}, {0x10233ad80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x10384f728?, 0x103825fa0?}}, {0x102362488, 0x14000482000}, {0x14000f22000, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102366a38?, 0x1400000fc68?}, {0x10197ffba?, 0x0?}}, {0x1019be204, 0x28}, {0x140001360e0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 103 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) Other Goroutines goroutine 25 [running] github.com/onsi/ginkgo/v2/internal.extractRunningGoroutines() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:181 github.com/onsi/ginkgo/v2/internal.NewProgressReport(_, {{0x14000a02000, 0x2, 0x2}, {0x1400029c230, 0x2, 0x2}, {0x140005b1d40, 0x2, 0x2}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:75 github.com/onsi/ginkgo/v2/internal.(*Suite).generateProgressReport(_, _) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:381 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode(_, {0xdc, 0x4, {0x101990e1e, 0x11}, 0x14000068d70, {{0x102a698e1, 0x55}, 0x6b, {0x0, ...}, ...}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:980 github.com/onsi/ginkgo/v2/internal.(*group).attemptSpec(0x14000dc17f0, 0x1, {{0x140005b3208?, 0x1400029c230?, 0x2?}, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:199 github.com/onsi/ginkgo/v2/internal.(*group).run(0x14000dc17f0, {0x140008906c8?, 0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:349 github.com/onsi/ginkgo/v2/internal.(*Suite).runSpecs(0x14000013508, {0x10197cf2b, 0x5}, {0x10384f728, 0x0, 0x0}, {0x1400005a0f4, 0x3d}, 0x0, {0x14000092c08, ...}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:489 github.com/onsi/ginkgo/v2/internal.(*Suite).Run(_, {_, _}, {_, _, _}, {_, _}, _, {_, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:130 github.com/onsi/ginkgo/v2.RunSpecs({0x10233a0a0, 0x14000432fc0}, {0x10197cf2b, 0x5}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:300 github.com/deckhouse/virtualization/tests/e2e.TestE2E(0x14000432fc0) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:76 testing.tRunner(0x14000432fc0, 0x102329c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.(*T).Run in goroutine 1 /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1997 goroutine 1 [chan receive, 8 minutes] testing.(*T).Run(0x14000432e00, {0x10197f99a?, 0x14000519b38?}, 0x102329c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2005 testing.runTests.func1(0x14000432e00) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2477 testing.tRunner(0x14000432e00, 0x14000519c68) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.runTests(0x1400000ec60, {0x1037d4fd0, 0x1, 0x1}, {0x14000293a80?, 0xf?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2475 testing.(*M).Run(0x14000418500) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2337 main.main() _testmain.go:45 goroutine 32 [IO wait] internal/poll.runtime_pollWait(0x12fba5e00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140001d2f80?, 0x140001e7500?, 0x0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140001d2f80, {0x140001e7500, 0x3500, 0x3500}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 net.(*netFD).Read(0x140001d2f80, {0x140001e7500?, 0x140006c3818?, 0x1001f2c30?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/net/fd_posix.go:68 net.(*conn).Read(0x14000126058, {0x140001e7500?, 0x103001e7500?, 0x12f9ea6c0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/net/net.go:196 crypto/tls.(*atLeastReader).Read(0x14000404180, {0x140001e7500?, 0x140006c38d8?, 0x100507ed4?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:816 bytes.(*Buffer).ReadFrom(0x140002ad7a8, {0x10233bde0, 0x14000404180}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bytes/buffer.go:217 crypto/tls.(*Conn).readFromUntil(0x140002ad508, {0x12fbab000, 0x14000126058}, 0x140006c3980?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:838 crypto/tls.(*Conn).readRecordOrCCS(0x140002ad508, 0x0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:627 crypto/tls.(*Conn).readRecord(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:589 crypto/tls.(*Conn).Read(0x140002ad508, {0x14000639000, 0x1000, 0x1400007ac08?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:1392 bufio.(*Reader).Read(0x1400063a000, {0x140002a44a0, 0x9, 0x1001dfee8?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:245 io.ReadAtLeast({0x10233a3c0, 0x1400063a000}, {0x140002a44a0, 0x9, 0x9}, 0x9) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:335 io.ReadFull(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:354 golang.org/x/net/http2.readFrameHeader({0x140002a44a0, 0x9, 0x140006c3d40?}, {0x10233a3c0?, 0x1400063a000?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/frame.go:242 golang.org/x/net/http2.(*Framer).ReadFrame(0x140002a4460) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/frame.go:506 golang.org/x/net/http2.(*clientConnReadLoop).run(0x140006c3f98) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2258 golang.org/x/net/http2.(*ClientConn).readLoop(0x140005828c0) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2127 golang.org/x/net/http2.(*Transport).newClientConn in goroutine 31 /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:912 goroutine 146 [select, 8 minutes] golang.org/x/net/http2.(*clientStream).writeRequest(0x140005f4c00, 0x14000039b80, 0x0) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1570 golang.org/x/net/http2.(*clientStream).doRequest(0x140005f4c00, 0x0?, 0x0?) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1431 golang.org/x/net/http2.(*ClientConn).roundTrip in goroutine 12 /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1336 goroutine 46 [select] github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts.func2(0x0?) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:131 github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts in goroutine 25 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:128 goroutine 11 [syscall] os/signal.signal_recv() /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sigqueue.go:149 os/signal.loop() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal_unix.go:23 os/signal.Notify.func1.1 in goroutine 25 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal.go:152 goroutine 109 [select, 8 minutes] os/exec.(*Cmd).watchCtx(0x14000171680, 0x140006e89a0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 149 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 66 [select, 8 minutes] github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal.func1() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:32 github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal in goroutine 25 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:30 goroutine 108 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12fba5600, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000075440?, 0x14000f92000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000075440, {0x14000f92000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140001262c0, {0x14000f92000?, 0x140004af5a8?, 0x1001efa18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10233ad80, 0x14000720200}, {0x102339ec8, 0x14000f90000}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1400000fc01?, {0x10233ad80, 0x14000720200}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037d5230?, {0x10233ad80?, 0x14000720200?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10233ad80, 0x14000720200}, {0x102339fc0, 0x140001262c0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400026eff8?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 149 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 70 [select, 8 minutes] golang.org/x/net/http2.(*clientStream).writeRequest(0x14000171c80, 0x140003d7900, 0x0) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1570 golang.org/x/net/http2.(*clientStream).doRequest(0x14000171c80, 0x100261b64?, 0x140004a6000?) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1431 golang.org/x/net/http2.(*ClientConn).roundTrip in goroutine 12 /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1336 goroutine 151 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12fba6400, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140005aa360?, 0x14000ddc000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140005aa360, {0x14000ddc000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000f00028, {0x14000ddc000?, 0x140008205a8?, 0x1001efa18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10233ad80, 0x14000470a40}, {0x102339ec8, 0x1400048c130}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1400000fc01?, {0x10233ad80, 0x14000470a40}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037d5230?, {0x10233ad80?, 0x14000470a40?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10233ad80, 0x14000470a40}, {0x102339fc0, 0x14000f00028}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400026eff8?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 150 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 107 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12fba6000, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000074420?, 0x14000f88000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000074420, {0x14000f88000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140001262a0, {0x14000f88000?, 0x140004adda8?, 0x1001efa18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10233ad80, 0x140007201c0}, {0x102339ec8, 0x14000428d20}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x140004ade01?, {0x10233ad80, 0x140007201c0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037d5230?, {0x10233ad80?, 0x140007201c0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10233ad80, 0x140007201c0}, {0x102339fc0, 0x140001262a0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400011ae70?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 149 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 153 [select, 8 minutes] os/exec.(*Cmd).watchCtx(0x14000f24000, 0x14000f3e070) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 150 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 152 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12fba5c00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140005aa420?, 0x14000ece000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140005aa420, {0x14000ece000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000f00040, {0x14000ece000?, 0x140004b2da8?, 0x1001efa18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10233ad80, 0x14000470bc0}, {0x102339ec8, 0x14000594140}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1400011b401?, {0x10233ad80, 0x14000470bc0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037d5230?, {0x10233ad80?, 0x14000470bc0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10233ad80, 0x14000470bc0}, {0x102339fc0, 0x14000f00040}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140005f4480?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 150 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 There were additional failures detected after the initial failure. These are visible in the timeline + > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/12/25 21:51:32.188 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 @ 10/12/25 21:51:32.188 [INTERRUPTED] Interrupted by User In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/12/25 21:59:01.568 This is the Progress Report generated when the interrupt was received: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 7m29.366s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [It] (Node Runtime: 7m29.366s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 At [By Step] VDs should be in Ready phase (Step Runtime: 7m29.366s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 Spec Goroutine goroutine 103 [sync.WaitGroup.Wait, 8 minutes] sync.runtime_SemacquireWaitGroup(0x10022cb30?, 0xd0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x1400026f000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000292040, 0x2, 0x101ca5e58?}, {0x1019be204, 0x28}, {{0x0, 0x0, 0x0}, 0x14000796b70, {0x140006d9f50, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1019be204, 0x28}, {{0x0, 0x0, 0x0}, 0x14000796b70, {0x140006d9f50, 0x25}, {0x140006c9a40, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1019be204, 0x28}, {0x10197ceef, 0x5}, {{0x0, 0x0, 0x0}, 0x14000796b70, {0x140006d9f50, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 25 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 149 [syscall, 8 minutes] syscall.syscall6(0x100e60620?, 0x12f9e7060?, 0x10459cf30?, 0x90?, 0x14000273808?, 0x14000208510?, 0x140009baa58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140009baa88?, 0x1002ae6fc?, 0x90?, 0x1022aa120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029c770?, 0x140009baac4, 0x14000e60620?, 0x1400029c700?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000720240) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000e47408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000171680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000171680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x102341e88?, 0x1400029c620?}}, {0x102362488?, 0x1400029c620?}, {0x1400073a540?, 0xc2331d0c5aaf8ee8?}, {0x10233ad80, 0x140007201c0}, {0x10233ad80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x10384f728?, 0x103825fa0?}}, {0x102362488, 0x1400029c620}, {0x1400073a540, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102366a38?, 0x1400000fc68?}, {0x10197ffba?, 0x140004af698?}}, {0x1019be204, 0x28}, {0x140001360c0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 103 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 147 [sync.Cond.Wait, 8 minutes] sync.runtime_notifyListWait(0x140005f4c48, 0x8b) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140005f4c38) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140005f4c30, {0x14000879000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400076d440?}, {0x14000879000?, 0x101e15b40?, 0x14000e63340?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140005aa840) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140005aa840) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000613908) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000613908) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000613908, {0x14000cba963, 0xe569d, 0x100371c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x14000176848, {0x14000cba963, 0xe569d, 0xe569d}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000cba963?, 0x14000e669b8?, 0x14000cba814?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x140009bfda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x1400059a210, {0x12fdcba60, 0x140006ba420}, {0x10233b5a0, 0x14000075320}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 12 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 49 [sync.Cond.Wait, 8 minutes] sync.runtime_notifyListWait(0x14000171cc8, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x14000171cb8) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x14000171cb0, {0x14000878000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400076d440?}, {0x14000878000?, 0x101e15b40?, 0x14000e188f0?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140005aa780) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140005aa780) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000612608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000612608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000612608, {0x14000ba3174, 0xfce8c, 0x100371c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x140001762c8, {0x14000ba3174, 0xfce8c, 0xfce8c}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000ba3174?, 0x14000e1a778?, 0x14000ba30f2?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x14000901da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x1400059a210, {0x12fdcba60, 0x1400076d230}, {0x10233b5a0, 0x14000075320}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 12 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 150 [syscall, 8 minutes] syscall.syscall6(0x10026f040?, 0x12fbbea38?, 0x10459ca78?, 0x90?, 0x14000790008?, 0x14000f20090?, 0x14000606a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000606a88?, 0x1002ae6fc?, 0x90?, 0x1022aa120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140004821c0?, 0x14000606ac4, 0x1400026f040?, 0x14000482150?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000470c40) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000f28408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000f24000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000f24000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x102341e88?, 0x14000482000?}}, {0x102362488?, 0x14000482000?}, {0x14000f22000?, 0xc2331d0c5aafa658?}, {0x10233ad80, 0x14000470a40}, {0x10233ad80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x10384f728?, 0x103825fa0?}}, {0x102362488, 0x14000482000}, {0x14000f22000, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102366a38?, 0x1400000fc68?}, {0x10197ffba?, 0x0?}}, {0x1019be204, 0x28}, {0x140001360e0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 103 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) Other Goroutines goroutine 25 [running] github.com/onsi/ginkgo/v2/internal.extractRunningGoroutines() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:181 github.com/onsi/ginkgo/v2/internal.NewProgressReport(_, {{0x14000a02000, 0x2, 0x2}, {0x1400029c230, 0x2, 0x2}, {0x140005b1d40, 0x2, 0x2}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:75 github.com/onsi/ginkgo/v2/internal.(*Suite).generateProgressReport(_, _) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:381 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode(_, {0xdc, 0x4, {0x101990e1e, 0x11}, 0x14000068d70, {{0x102a698e1, 0x55}, 0x6b, {0x0, ...}, ...}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:980 github.com/onsi/ginkgo/v2/internal.(*group).attemptSpec(0x14000dc17f0, 0x1, {{0x140005b3208?, 0x1400029c230?, 0x2?}, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:199 github.com/onsi/ginkgo/v2/internal.(*group).run(0x14000dc17f0, {0x140008906c8?, 0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:349 github.com/onsi/ginkgo/v2/internal.(*Suite).runSpecs(0x14000013508, {0x10197cf2b, 0x5}, {0x10384f728, 0x0, 0x0}, {0x1400005a0f4, 0x3d}, 0x0, {0x14000092c08, ...}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:489 github.com/onsi/ginkgo/v2/internal.(*Suite).Run(_, {_, _}, {_, _, _}, {_, _}, _, {_, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:130 github.com/onsi/ginkgo/v2.RunSpecs({0x10233a0a0, 0x14000432fc0}, {0x10197cf2b, 0x5}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:300 github.com/deckhouse/virtualization/tests/e2e.TestE2E(0x14000432fc0) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:76 testing.tRunner(0x14000432fc0, 0x102329c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.(*T).Run in goroutine 1 /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1997 goroutine 1 [chan receive, 8 minutes] testing.(*T).Run(0x14000432e00, {0x10197f99a?, 0x14000519b38?}, 0x102329c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2005 testing.runTests.func1(0x14000432e00) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2477 testing.tRunner(0x14000432e00, 0x14000519c68) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.runTests(0x1400000ec60, {0x1037d4fd0, 0x1, 0x1}, {0x14000293a80?, 0xf?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2475 testing.(*M).Run(0x14000418500) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2337 main.main() _testmain.go:45 goroutine 32 [IO wait] internal/poll.runtime_pollWait(0x12fba5e00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140001d2f80?, 0x140001e7500?, 0x0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140001d2f80, {0x140001e7500, 0x3500, 0x3500}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 net.(*netFD).Read(0x140001d2f80, {0x140001e7500?, 0x140006c3818?, 0x1001f2c30?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/net/fd_posix.go:68 net.(*conn).Read(0x14000126058, {0x140001e7500?, 0x103001e7500?, 0x12f9ea6c0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/net/net.go:196 crypto/tls.(*atLeastReader).Read(0x14000404180, {0x140001e7500?, 0x140006c38d8?, 0x100507ed4?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:816 bytes.(*Buffer).ReadFrom(0x140002ad7a8, {0x10233bde0, 0x14000404180}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bytes/buffer.go:217 crypto/tls.(*Conn).readFromUntil(0x140002ad508, {0x12fbab000, 0x14000126058}, 0x140006c3980?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:838 crypto/tls.(*Conn).readRecordOrCCS(0x140002ad508, 0x0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:627 crypto/tls.(*Conn).readRecord(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:589 crypto/tls.(*Conn).Read(0x140002ad508, {0x14000639000, 0x1000, 0x1400007ac08?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:1392 bufio.(*Reader).Read(0x1400063a000, {0x140002a44a0, 0x9, 0x1001dfee8?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:245 io.ReadAtLeast({0x10233a3c0, 0x1400063a000}, {0x140002a44a0, 0x9, 0x9}, 0x9) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:335 io.ReadFull(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:354 golang.org/x/net/http2.readFrameHeader({0x140002a44a0, 0x9, 0x140006c3d40?}, {0x10233a3c0?, 0x1400063a000?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/frame.go:242 golang.org/x/net/http2.(*Framer).ReadFrame(0x140002a4460) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/frame.go:506 golang.org/x/net/http2.(*clientConnReadLoop).run(0x140006c3f98) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2258 golang.org/x/net/http2.(*ClientConn).readLoop(0x140005828c0) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2127 golang.org/x/net/http2.(*Transport).newClientConn in goroutine 31 /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:912 goroutine 146 [select, 8 minutes] golang.org/x/net/http2.(*clientStream).writeRequest(0x140005f4c00, 0x14000039b80, 0x0) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1570 golang.org/x/net/http2.(*clientStream).doRequest(0x140005f4c00, 0x0?, 0x0?) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1431 golang.org/x/net/http2.(*ClientConn).roundTrip in goroutine 12 /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1336 goroutine 46 [select] github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts.func2(0x0?) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:131 github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts in goroutine 25 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:128 goroutine 11 [syscall] os/signal.signal_recv() /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sigqueue.go:149 os/signal.loop() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal_unix.go:23 os/signal.Notify.func1.1 in goroutine 25 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal.go:152 goroutine 109 [select, 8 minutes] os/exec.(*Cmd).watchCtx(0x14000171680, 0x140006e89a0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 149 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 66 [select, 8 minutes] github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal.func1() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:32 github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal in goroutine 25 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:30 goroutine 108 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12fba5600, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000075440?, 0x14000f92000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000075440, {0x14000f92000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140001262c0, {0x14000f92000?, 0x140004af5a8?, 0x1001efa18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10233ad80, 0x14000720200}, {0x102339ec8, 0x14000f90000}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1400000fc01?, {0x10233ad80, 0x14000720200}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037d5230?, {0x10233ad80?, 0x14000720200?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10233ad80, 0x14000720200}, {0x102339fc0, 0x140001262c0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400026eff8?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 149 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 70 [select, 8 minutes] golang.org/x/net/http2.(*clientStream).writeRequest(0x14000171c80, 0x140003d7900, 0x0) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1570 golang.org/x/net/http2.(*clientStream).doRequest(0x14000171c80, 0x100261b64?, 0x140004a6000?) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1431 golang.org/x/net/http2.(*ClientConn).roundTrip in goroutine 12 /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1336 goroutine 151 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12fba6400, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140005aa360?, 0x14000ddc000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140005aa360, {0x14000ddc000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000f00028, {0x14000ddc000?, 0x140008205a8?, 0x1001efa18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10233ad80, 0x14000470a40}, {0x102339ec8, 0x1400048c130}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1400000fc01?, {0x10233ad80, 0x14000470a40}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037d5230?, {0x10233ad80?, 0x14000470a40?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10233ad80, 0x14000470a40}, {0x102339fc0, 0x14000f00028}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400026eff8?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 150 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 107 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12fba6000, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000074420?, 0x14000f88000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000074420, {0x14000f88000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140001262a0, {0x14000f88000?, 0x140004adda8?, 0x1001efa18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10233ad80, 0x140007201c0}, {0x102339ec8, 0x14000428d20}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x140004ade01?, {0x10233ad80, 0x140007201c0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037d5230?, {0x10233ad80?, 0x140007201c0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10233ad80, 0x140007201c0}, {0x102339fc0, 0x140001262a0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400011ae70?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 149 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 153 [select, 8 minutes] os/exec.(*Cmd).watchCtx(0x14000f24000, 0x14000f3e070) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 150 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 152 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12fba5c00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140005aa420?, 0x14000ece000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140005aa420, {0x14000ece000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000f00040, {0x14000ece000?, 0x140004b2da8?, 0x1001efa18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10233ad80, 0x14000470bc0}, {0x102339ec8, 0x14000594140}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1400011b401?, {0x10233ad80, 0x14000470bc0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037d5230?, {0x10233ad80?, 0x14000470bc0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10233ad80, 0x14000470bc0}, {0x102339fc0, 0x14000f00040}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140005f4480?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 150 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 Interrupted by User First interrupt received; Ginkgo will run any cleanup and reporting nodes but will skip all remaining specs. Interrupt again to skip cleanup. Here's a current progress report: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 7m29.366s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [It] (Node Runtime: 7m29.366s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 At [By Step] VDs should be in Ready phase (Step Runtime: 7m29.366s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 Spec Goroutine goroutine 103 [sync.WaitGroup.Wait, 8 minutes] sync.runtime_SemacquireWaitGroup(0x10022cb30?, 0xd0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x1400026f000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000292040, 0x2, 0x101ca5e58?}, {0x1019be204, 0x28}, {{0x0, 0x0, 0x0}, 0x14000796b70, {0x140006d9f50, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1019be204, 0x28}, {{0x0, 0x0, 0x0}, 0x14000796b70, {0x140006d9f50, 0x25}, {0x140006c9a40, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1019be204, 0x28}, {0x10197ceef, 0x5}, {{0x0, 0x0, 0x0}, 0x14000796b70, {0x140006d9f50, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 25 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 149 [syscall, 8 minutes] syscall.syscall6(0x100e60620?, 0x12f9e7060?, 0x10459cf30?, 0x90?, 0x14000273808?, 0x14000208510?, 0x140009baa58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140009baa88?, 0x1002ae6fc?, 0x90?, 0x1022aa120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029c770?, 0x140009baac4, 0x14000e60620?, 0x1400029c700?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000720240) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000e47408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000171680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000171680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x102341e88?, 0x1400029c620?}}, {0x102362488?, 0x1400029c620?}, {0x1400073a540?, 0xc2331d0c5aaf8ee8?}, {0x10233ad80, 0x140007201c0}, {0x10233ad80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x10384f728?, 0x103825fa0?}}, {0x102362488, 0x1400029c620}, {0x1400073a540, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102366a38?, 0x1400000fc68?}, {0x10197ffba?, 0x140004af698?}}, {0x1019be204, 0x28}, {0x140001360c0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 103 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 147 [sync.Cond.Wait, 8 minutes] sync.runtime_notifyListWait(0x140005f4c48, 0x8b) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140005f4c38) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140005f4c30, {0x14000879000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400076d440?}, {0x14000879000?, 0x101e15b40?, 0x14000e63340?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140005aa840) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140005aa840) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000613908) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000613908) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000613908, {0x14000cba963, 0xe569d, 0x100371c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x14000176848, {0x14000cba963, 0xe569d, 0xe569d}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000cba963?, 0x14000e669b8?, 0x14000cba814?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x140009bfda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x1400059a210, {0x12fdcba60, 0x140006ba420}, {0x10233b5a0, 0x14000075320}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 12 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 49 [sync.Cond.Wait, 8 minutes] sync.runtime_notifyListWait(0x14000171cc8, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x14000171cb8) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x14000171cb0, {0x14000878000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400076d440?}, {0x14000878000?, 0x101e15b40?, 0x14000e188f0?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140005aa780) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140005aa780) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000612608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000612608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000612608, {0x14000ba3174, 0xfce8c, 0x100371c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x140001762c8, {0x14000ba3174, 0xfce8c, 0xfce8c}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000ba3174?, 0x14000e1a778?, 0x14000ba30f2?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x14000901da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x1400059a210, {0x12fdcba60, 0x1400076d230}, {0x10233b5a0, 0x14000075320}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 12 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 150 [syscall, 8 minutes] syscall.syscall6(0x10026f040?, 0x12fbbea38?, 0x10459ca78?, 0x90?, 0x14000790008?, 0x14000f20090?, 0x14000606a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000606a88?, 0x1002ae6fc?, 0x90?, 0x1022aa120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140004821c0?, 0x14000606ac4, 0x1400026f040?, 0x14000482150?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000470c40) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000f28408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000f24000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000f24000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x102341e88?, 0x14000482000?}}, {0x102362488?, 0x14000482000?}, {0x14000f22000?, 0xc2331d0c5aafa658?}, {0x10233ad80, 0x14000470a40}, {0x10233ad80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x10384f728?, 0x103825fa0?}}, {0x102362488, 0x14000482000}, {0x14000f22000, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102366a38?, 0x1400000fc68?}, {0x10197ffba?, 0x0?}}, {0x1019be204, 0x28}, {0x140001360e0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 103 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/12/25 21:59:01.571 (7m29.368s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 21:59:01.571 [INTERRUPTED] Interrupted by User In [AfterEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 21:59:04.806 This is the Progress Report generated when the interrupt was received: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 7m32.604s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [AfterEach] (Node Runtime: 3.235s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 Spec Goroutine goroutine 103 [sync.WaitGroup.Wait, 8 minutes] sync.runtime_SemacquireWaitGroup(0x10022cb30?, 0xd0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x1400026f000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000292040, 0x2, 0x101ca5e58?}, {0x1019be204, 0x28}, {{0x0, 0x0, 0x0}, 0x14000796b70, {0x140006d9f50, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1019be204, 0x28}, {{0x0, 0x0, 0x0}, 0x14000796b70, {0x140006d9f50, 0x25}, {0x140006c9a40, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1019be204, 0x28}, {0x10197ceef, 0x5}, {{0x0, 0x0, 0x0}, 0x14000796b70, {0x140006d9f50, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 25 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 149 [syscall, 8 minutes] syscall.syscall6(0x100e60620?, 0x12f9e7060?, 0x10459cf30?, 0x90?, 0x14000273808?, 0x14000208510?, 0x140009baa58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140009baa88?, 0x1002ae6fc?, 0x90?, 0x1022aa120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029c770?, 0x140009baac4, 0x14000e60620?, 0x1400029c700?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000720240) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000e47408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000171680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000171680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x102341e88?, 0x1400029c620?}}, {0x102362488?, 0x1400029c620?}, {0x1400073a540?, 0xc2331d0c5aaf8ee8?}, {0x10233ad80, 0x140007201c0}, {0x10233ad80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x10384f728?, 0x103825fa0?}}, {0x102362488, 0x1400029c620}, {0x1400073a540, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102366a38?, 0x1400000fc68?}, {0x10197ffba?, 0x140004af698?}}, {0x1019be204, 0x28}, {0x140001360c0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 103 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 147 [sync.Cond.Wait, 8 minutes] sync.runtime_notifyListWait(0x140005f4c48, 0x8b) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140005f4c38) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140005f4c30, {0x14000879000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400076d440?}, {0x14000879000?, 0x101e15b40?, 0x14000e63340?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140005aa840) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140005aa840) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000613908) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000613908) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000613908, {0x14000cba963, 0xe569d, 0x100371c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x14000176848, {0x14000cba963, 0xe569d, 0xe569d}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000cba963?, 0x14000e669b8?, 0x14000cba814?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x140009bfda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x1400059a210, {0x12fdcba60, 0x140006ba420}, {0x10233b5a0, 0x14000075320}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 12 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 49 [sync.Cond.Wait, 8 minutes] sync.runtime_notifyListWait(0x14000171cc8, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x14000171cb8) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x14000171cb0, {0x14000878000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400076d440?}, {0x14000878000?, 0x101e15b40?, 0x14000e188f0?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140005aa780) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140005aa780) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000612608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000612608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000612608, {0x14000ba3174, 0xfce8c, 0x100371c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x140001762c8, {0x14000ba3174, 0xfce8c, 0xfce8c}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000ba3174?, 0x14000e1a778?, 0x14000ba30f2?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x14000901da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x1400059a210, {0x12fdcba60, 0x1400076d230}, {0x10233b5a0, 0x14000075320}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 12 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 150 [syscall, 8 minutes] syscall.syscall6(0x10026f040?, 0x12fbbea38?, 0x10459ca78?, 0x90?, 0x14000790008?, 0x14000f20090?, 0x14000606a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000606a88?, 0x1002ae6fc?, 0x90?, 0x1022aa120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140004821c0?, 0x14000606ac4, 0x1400026f040?, 0x14000482150?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000470c40) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000f28408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000f24000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000f24000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x102341e88?, 0x14000482000?}}, {0x102362488?, 0x14000482000?}, {0x14000f22000?, 0xc2331d0c5aafa658?}, {0x10233ad80, 0x14000470a40}, {0x10233ad80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x10384f728?, 0x103825fa0?}}, {0x102362488, 0x14000482000}, {0x14000f22000, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102366a38?, 0x1400000fc68?}, {0x10197ffba?, 0x0?}}, {0x1019be204, 0x28}, {0x140001360e0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 103 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 194 [syscall] syscall.syscall6(0x10090f9a0?, 0x12fbbea38?, 0x10459cf30?, 0x90?, 0x14000100008?, 0x14000f20b40?, 0x14000902b08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000902b38?, 0x1002ae6fc?, 0x90?, 0x1022aa120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000482770?, 0x14000902b74, 0x1400090f9a0?, 0x14000482700?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140007846c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000051008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140005ca000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140005ca000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x102341e88?, 0x14000482620?}}, {0x102362488?, 0x14000482620?}, {0x14000924840?, 0xc2331c84e20b9380?}, {0x10233ad80, 0x140007845c0}, {0x10233ad80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x10384f728?, 0x103825fa0?}}, {0x102362488, 0x14000482620}, {0x14000924840, 0xb0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.Get({{0x102366a38?, 0x1400000fc68?}, {0x10197ffba?, 0x140007a3c80?}}, {0x1019d72e2, 0x31}, {{0x0, 0x0, 0x0}, 0x0, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:208 | ctx, cancel := context.WithTimeout(context.Background(), MediumTimeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.SaveTestCaseResources(0x14000796b70, {0x14000719a28, 0x11}, {0x140006d9f50, 0x25}, {0x10197c370, 0x4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:765 | errorFileName := fmt.Sprintf("%s/e2e_failed__%s__%s_error.txt", dumpPath, labels["testcase"], additional) | > cmdr := kubectl.Get( | "virtualization,cvi,vmc,intvirt,pod,volumesnapshot", | kc.GetOptions{ > github.com/deckhouse/virtualization/tests/e2e.SaveTestCaseDump(0x14000796b70, {0x101990e1e, 0x11}, {0x140006d9f50, 0x25}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:757 | } | > SaveTestCaseResources(labels, additional, namespace, tmpDir) | SavePodLogsAndDescriptions(labels, additional, namespace, tmpDir) | } > github.com/deckhouse/virtualization/tests/e2e.init.func12.2() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:68 | AfterEach(func() { | if CurrentSpecReport().Failed() { > SaveTestCaseDump(testCaseLabel, CurrentSpecReport().LeafNodeText, ns) | } | }) github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 25 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Other Goroutines goroutine 25 [running] github.com/onsi/ginkgo/v2/internal.extractRunningGoroutines() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:181 github.com/onsi/ginkgo/v2/internal.NewProgressReport(_, {{0x14000a02000, 0x2, 0x2}, {0x1400029c230, 0x2, 0x2}, {0x140005b1d40, 0x2, 0x2}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:75 github.com/onsi/ginkgo/v2/internal.(*Suite).generateProgressReport(_, _) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:381 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode(_, {0xd6, 0x20, {0x0, 0x0}, 0x14000068ba0, {{0x102a698e1, 0x55}, 0x42, {0x0, ...}, ...}, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:980 github.com/onsi/ginkgo/v2/internal.(*group).attemptSpec(0x14000dc17f0, 0x1, {{0x140005b3208?, 0x1400029c230?, 0x2?}, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:286 github.com/onsi/ginkgo/v2/internal.(*group).run(0x14000dc17f0, {0x140008906c8?, 0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/group.go:349 github.com/onsi/ginkgo/v2/internal.(*Suite).runSpecs(0x14000013508, {0x10197cf2b, 0x5}, {0x10384f728, 0x0, 0x0}, {0x1400005a0f4, 0x3d}, 0x0, {0x14000092c08, ...}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:489 github.com/onsi/ginkgo/v2/internal.(*Suite).Run(_, {_, _}, {_, _, _}, {_, _}, _, {_, ...}, ...) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:130 github.com/onsi/ginkgo/v2.RunSpecs({0x10233a0a0, 0x14000432fc0}, {0x10197cf2b, 0x5}, {0x0, 0x0, 0x0}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/core_dsl.go:300 github.com/deckhouse/virtualization/tests/e2e.TestE2E(0x14000432fc0) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:76 testing.tRunner(0x14000432fc0, 0x102329c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.(*T).Run in goroutine 1 /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1997 goroutine 1 [chan receive, 8 minutes] testing.(*T).Run(0x14000432e00, {0x10197f99a?, 0x14000519b38?}, 0x102329c50) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2005 testing.runTests.func1(0x14000432e00) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2477 testing.tRunner(0x14000432e00, 0x14000519c68) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:1934 testing.runTests(0x1400000ec60, {0x1037d4fd0, 0x1, 0x1}, {0x14000293a80?, 0xf?, 0x0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2475 testing.(*M).Run(0x14000418500) /opt/homebrew/Cellar/go/1.25.1/libexec/src/testing/testing.go:2337 main.main() _testmain.go:45 goroutine 32 [IO wait] internal/poll.runtime_pollWait(0x12fba5e00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140001d2f80?, 0x140001e7500?, 0x0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140001d2f80, {0x140001e7500, 0x3500, 0x3500}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 net.(*netFD).Read(0x140001d2f80, {0x140001e7500?, 0x140006c3818?, 0x1001f2c30?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/net/fd_posix.go:68 net.(*conn).Read(0x14000126058, {0x140001e7500?, 0x103001e7500?, 0x12f9ea6c0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/net/net.go:196 crypto/tls.(*atLeastReader).Read(0x14000404180, {0x140001e7500?, 0x140006c38d8?, 0x100507ed4?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:816 bytes.(*Buffer).ReadFrom(0x140002ad7a8, {0x10233bde0, 0x14000404180}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bytes/buffer.go:217 crypto/tls.(*Conn).readFromUntil(0x140002ad508, {0x12fbab000, 0x14000126058}, 0x140006c3980?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:838 crypto/tls.(*Conn).readRecordOrCCS(0x140002ad508, 0x0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:627 crypto/tls.(*Conn).readRecord(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:589 crypto/tls.(*Conn).Read(0x140002ad508, {0x14000639000, 0x1000, 0x1400007ac08?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/crypto/tls/conn.go:1392 bufio.(*Reader).Read(0x1400063a000, {0x140002a44a0, 0x9, 0x1001dfee8?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:245 io.ReadAtLeast({0x10233a3c0, 0x1400063a000}, {0x140002a44a0, 0x9, 0x9}, 0x9) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:335 io.ReadFull(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:354 golang.org/x/net/http2.readFrameHeader({0x140002a44a0, 0x9, 0x140006c3d40?}, {0x10233a3c0?, 0x1400063a000?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/frame.go:242 golang.org/x/net/http2.(*Framer).ReadFrame(0x140002a4460) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/frame.go:506 golang.org/x/net/http2.(*clientConnReadLoop).run(0x140006c3f98) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2258 golang.org/x/net/http2.(*ClientConn).readLoop(0x140005828c0) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2127 golang.org/x/net/http2.(*Transport).newClientConn in goroutine 31 /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:912 goroutine 146 [select, 8 minutes] golang.org/x/net/http2.(*clientStream).writeRequest(0x140005f4c00, 0x14000039b80, 0x0) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1570 golang.org/x/net/http2.(*clientStream).doRequest(0x140005f4c00, 0x0?, 0x0?) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1431 golang.org/x/net/http2.(*ClientConn).roundTrip in goroutine 12 /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1336 goroutine 46 [select] github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts.func2(0x0?) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:131 github.com/onsi/ginkgo/v2/internal/interrupt_handler.(*InterruptHandler).registerForInterrupts in goroutine 25 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/interrupt_handler/interrupt_handler.go:128 goroutine 11 [syscall] os/signal.signal_recv() /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sigqueue.go:149 os/signal.loop() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal_unix.go:23 os/signal.Notify.func1.1 in goroutine 25 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/signal/signal.go:152 goroutine 109 [select, 8 minutes] os/exec.(*Cmd).watchCtx(0x14000171680, 0x140006e89a0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 149 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 66 [select, 8 minutes] github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal.func1() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:32 github.com/onsi/ginkgo/v2/internal.RegisterForProgressSignal in goroutine 25 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/progress_report.go:30 goroutine 108 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12fba5600, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000075440?, 0x14000f92000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000075440, {0x14000f92000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140001262c0, {0x14000f92000?, 0x140004af5a8?, 0x1001efa18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10233ad80, 0x14000720200}, {0x102339ec8, 0x14000f90000}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1400000fc01?, {0x10233ad80, 0x14000720200}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037d5230?, {0x10233ad80?, 0x14000720200?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10233ad80, 0x14000720200}, {0x102339fc0, 0x140001262c0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400026eff8?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 149 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 70 [select, 8 minutes] golang.org/x/net/http2.(*clientStream).writeRequest(0x14000171c80, 0x140003d7900, 0x0) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1570 golang.org/x/net/http2.(*clientStream).doRequest(0x14000171c80, 0x100261b64?, 0x140004a6000?) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1431 golang.org/x/net/http2.(*ClientConn).roundTrip in goroutine 12 /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:1336 goroutine 151 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12fba6400, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140005aa360?, 0x14000ddc000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140005aa360, {0x14000ddc000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000f00028, {0x14000ddc000?, 0x140008205a8?, 0x1001efa18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10233ad80, 0x14000470a40}, {0x102339ec8, 0x1400048c130}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1400000fc01?, {0x10233ad80, 0x14000470a40}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037d5230?, {0x10233ad80?, 0x14000470a40?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10233ad80, 0x14000470a40}, {0x102339fc0, 0x14000f00028}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400026eff8?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 150 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 107 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12fba6000, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x14000074420?, 0x14000f88000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x14000074420, {0x14000f88000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x140001262a0, {0x14000f88000?, 0x140004adda8?, 0x1001efa18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10233ad80, 0x140007201c0}, {0x102339ec8, 0x14000428d20}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x140004ade01?, {0x10233ad80, 0x140007201c0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037d5230?, {0x10233ad80?, 0x140007201c0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10233ad80, 0x140007201c0}, {0x102339fc0, 0x140001262a0}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400011ae70?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 149 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 153 [select, 8 minutes] os/exec.(*Cmd).watchCtx(0x14000f24000, 0x14000f3e070) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 150 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 goroutine 152 [IO wait, 8 minutes] internal/poll.runtime_pollWait(0x12fba5c00, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x140005aa420?, 0x14000ece000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x140005aa420, {0x14000ece000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x14000f00040, {0x14000ece000?, 0x140004b2da8?, 0x1001efa18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10233ad80, 0x14000470bc0}, {0x102339ec8, 0x14000594140}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1400011b401?, {0x10233ad80, 0x14000470bc0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037d5230?, {0x10233ad80?, 0x14000470bc0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10233ad80, 0x14000470bc0}, {0x102339fc0, 0x14000f00040}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x140005f4480?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 150 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 195 [IO wait] internal/poll.runtime_pollWait(0x12fbf9000, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x1400063a480?, 0x1400082c000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x1400063a480, {0x1400082c000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x1400048c088, {0x1400082c000?, 0x14000604da8?, 0x1001efa18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10233ad80, 0x140007845c0}, {0x102339ec8, 0x14000428000}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x10233ad80, 0x140007845c0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037d5230?, {0x10233ad80?, 0x140007845c0?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10233ad80, 0x140007845c0}, {0x102339fc0, 0x1400048c088}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x14000664008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 194 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 196 [IO wait] internal/poll.runtime_pollWait(0x12fba5800, 0x72) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/netpoll.go:351 internal/poll.(*pollDesc).wait(0x1400063a540?, 0x14000df4000?, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:84 internal/poll.(*pollDesc).waitRead(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0x1400063a540, {0x14000df4000, 0x8000, 0x8000}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/internal/poll/fd_unix.go:165 os.(*File).read(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:29 os.(*File).Read(0x1400048c0c8, {0x14000df4000?, 0x1400009cda8?, 0x1001efa18?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:144 io.copyBuffer({0x10233ad80, 0x14000784640}, {0x102339ec8, 0x14000f00000}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:429 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os.genericWriteTo(0x1?, {0x10233ad80, 0x14000784640}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:295 os.(*File).WriteTo(0x1037d5230?, {0x10233ad80?, 0x14000784640?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file.go:273 io.copyBuffer({0x10233ad80, 0x14000784640}, {0x102339fc0, 0x1400048c0c8}, {0x0, 0x0, 0x0}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:411 io.Copy(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/io/io.go:388 os/exec.(*Cmd).writerDescriptor.func1() /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:596 os/exec.(*Cmd).Start.func2(0x1400026e030?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:749 os/exec.(*Cmd).Start in goroutine 194 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:748 goroutine 197 [select] os/exec.(*Cmd).watchCtx(0x140005ca000, 0x14000f3f9d0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:789 os/exec.(*Cmd).Start in goroutine 194 /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:775 Interrupted by User Second interrupt received; Ginkgo will run any reporting nodes but will skip all remaining specs and cleanup nodes. Interrupt again to bail immediately. Here's a current progress report: VirtualMachineConnectivity When virtual disks are applied checks VDs phases (Spec Runtime: 7m32.604s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 In [AfterEach] (Node Runtime: 3.235s) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 Spec Goroutine goroutine 103 [sync.WaitGroup.Wait, 8 minutes] sync.runtime_SemacquireWaitGroup(0x10022cb30?, 0xd0?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:114 sync.(*WaitGroup).Wait(0x1400026f000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/waitgroup.go:206 > github.com/deckhouse/virtualization/tests/e2e.WaitResources({0x14000292040, 0x2, 0x101ca5e58?}, {0x1019be204, 0x28}, {{0x0, 0x0, 0x0}, 0x14000796b70, {0x140006d9f50, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:349 | }() | } > wg.Wait() | Expect(waitErr).To(BeEmpty(), "should observe resources in '%s' state before %s timeout", opts.For, opts.Timeout.String()) | } > github.com/deckhouse/virtualization/tests/e2e.WaitByLabel({0x1019be204, 0x28}, {{0x0, 0x0, 0x0}, 0x14000796b70, {0x140006d9f50, 0x25}, {0x140006c9a40, 0x20}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:308 | resources = strings.Split(res.StdOut(), " ") | } > WaitResources(resources, resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitPhaseByLabel({0x1019be204, 0x28}, {0x10197ceef, 0x5}, {{0x0, 0x0, 0x0}, 0x14000796b70, {0x140006d9f50, 0x25}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:290 | GinkgoHelper() | opts.For = fmt.Sprintf("'jsonpath={.status.phase}=%s'", phase) > WaitByLabel(resource, opts) | } | > github.com/deckhouse/virtualization/tests/e2e.init.func12.5.1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:109 | It("checks VDs phases", func() { | By(fmt.Sprintf("VDs should be in %s phase", PhaseReady)) > WaitPhaseByLabel(kc.ResourceVD, PhaseReady, kc.WaitOptions{ | Labels: testCaseLabel, | Namespace: ns, github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 25 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 Goroutines of Interest goroutine 149 [syscall, 8 minutes] syscall.syscall6(0x100e60620?, 0x12f9e7060?, 0x10459cf30?, 0x90?, 0x14000273808?, 0x14000208510?, 0x140009baa58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x140009baa88?, 0x1002ae6fc?, 0x90?, 0x1022aa120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x1400029c770?, 0x140009baac4, 0x14000e60620?, 0x1400029c700?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000720240) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000e47408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000171680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000171680) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x102341e88?, 0x1400029c620?}}, {0x102362488?, 0x1400029c620?}, {0x1400073a540?, 0xc2331d0c5aaf8ee8?}, {0x10233ad80, 0x140007201c0}, {0x10233ad80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x10384f728?, 0x103825fa0?}}, {0x102362488, 0x1400029c620}, {0x1400073a540, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102366a38?, 0x1400000fc68?}, {0x10197ffba?, 0x140004af698?}}, {0x1019be204, 0x28}, {0x140001360c0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 103 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 147 [sync.Cond.Wait, 8 minutes] sync.runtime_notifyListWait(0x140005f4c48, 0x8b) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x140005f4c38) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x140005f4c30, {0x14000879000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400076d440?}, {0x14000879000?, 0x101e15b40?, 0x14000e63340?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140005aa840) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140005aa840) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000613908) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000613908) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000613908, {0x14000cba963, 0xe569d, 0x100371c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x14000176848, {0x14000cba963, 0xe569d, 0xe569d}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000cba963?, 0x14000e669b8?, 0x14000cba814?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x140009bfda0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x1400059a210, {0x12fdcba60, 0x140006ba420}, {0x10233b5a0, 0x14000075320}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 12 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 49 [sync.Cond.Wait, 8 minutes] sync.runtime_notifyListWait(0x14000171cc8, 0x1) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sema.go:606 sync.(*Cond).Wait(0x14000171cb8) /opt/homebrew/Cellar/go/1.25.1/libexec/src/sync/cond.go:71 golang.org/x/net/http2.(*pipe).Read(0x14000171cb0, {0x14000878000, 0x1000, 0x1000}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/pipe.go:76 golang.org/x/net/http2.transportResponseBody.Read({0x1400076d440?}, {0x14000878000?, 0x101e15b40?, 0x14000e188f0?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:2560 bufio.(*Reader).fill(0x140005aa780) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:113 bufio.(*Reader).ReadByte(0x140005aa780) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/bufio.go:273 compress/flate.(*decompressor).moreBits(0x14000612608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:697 compress/flate.(*decompressor).nextBlock(0x14000612608) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:304 compress/flate.(*decompressor).Read(0x14000612608, {0x14000ba3174, 0xfce8c, 0x100371c74?}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/flate/inflate.go:348 compress/gzip.(*Reader).Read(0x140001762c8, {0x14000ba3174, 0xfce8c, 0xfce8c}) /opt/homebrew/Cellar/go/1.25.1/libexec/src/compress/gzip/gunzip.go:252 golang.org/x/net/http2.(*gzipReader).Read(0x0?, {0x14000ba3174?, 0x14000e1a778?, 0x14000ba30f2?}) /Users/antont/go/pkg/mod/golang.org/x/net@v0.39.0/http2/transport.go:3151 bufio.(*Scanner).Scan(0x14000901da0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/bufio/scan.go:219 > github.com/deckhouse/virtualization/tests/e2e/errlogger.(*LogStreamer).Stream(0x1400059a210, {0x12fdcba60, 0x1400076d230}, {0x10233b5a0, 0x14000075320}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/errlogger/errlogger.go:78 | num := 0 | > for scanner.Scan() { | rawEntry := scanner.Bytes() | > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:248 | defer l.wg.Done() | > n, err := logStreamer.Stream(readCloser, GinkgoWriter) | l.mu.Lock() | defer l.mu.Unlock() > github.com/deckhouse/virtualization/tests/e2e.(*logStreamer).Start in goroutine 12 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:245 | | l.wg.Add(1) > go func() { | defer l.wg.Done() | goroutine 150 [syscall, 8 minutes] syscall.syscall6(0x10026f040?, 0x12fbbea38?, 0x10459ca78?, 0x90?, 0x14000790008?, 0x14000f20090?, 0x14000606a58?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000606a88?, 0x1002ae6fc?, 0x90?, 0x1022aa120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x140004821c0?, 0x14000606ac4, 0x1400026f040?, 0x14000482150?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x14000470c40) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000f28408?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x14000f24000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x14000f24000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x102341e88?, 0x14000482000?}}, {0x102362488?, 0x14000482000?}, {0x14000f22000?, 0xc2331d0c5aafa658?}, {0x10233ad80, 0x14000470a40}, {0x10233ad80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x10384f728?, 0x103825fa0?}}, {0x102362488, 0x14000482000}, {0x14000f22000, 0xbf}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.WaitResource({{0x102366a38?, 0x1400000fc68?}, {0x10197ffba?, 0x0?}}, {0x1019be204, 0x28}, {0x140001360e0, 0x1f}, {{0x0, 0x0, ...}, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:260 | ctx, cancel := context.WithTimeout(context.Background(), timeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.WaitResources.func1() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:341 | go func() { | defer wg.Done() > res := kubectl.WaitResource(resource, name, waitOpts) | if res.Error() != nil { | mu.Lock() > github.com/deckhouse/virtualization/tests/e2e.WaitResources in goroutine 103 /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:339 | for _, name := range resources { | wg.Add(1) > go func() { | defer wg.Done() | res := kubectl.WaitResource(resource, name, waitOpts) goroutine 194 [syscall] syscall.syscall6(0x10090f9a0?, 0x12fbbea38?, 0x10459cf30?, 0x90?, 0x14000100008?, 0x14000f20b40?, 0x14000902b08?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/runtime/sys_darwin.go:60 syscall.wait4(0x14000902b38?, 0x1002ae6fc?, 0x90?, 0x1022aa120?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/zsyscall_darwin_arm64.go:44 syscall.Wait4(0x14000482770?, 0x14000902b74, 0x1400090f9a0?, 0x14000482700?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/syscall/syscall_bsd.go:144 os.(*Process).pidWait.func1(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:64 os.ignoringEINTR2[...](...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/file_posix.go:266 os.(*Process).pidWait(0x140007846c0) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:63 os.(*Process).wait(0x14000051008?) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec_unix.go:28 os.(*Process).Wait(...) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec.go:340 os/exec.(*Cmd).Wait(0x140005ca000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:922 os/exec.(*Cmd).Run(0x140005ca000) /opt/homebrew/Cellar/go/1.25.1/libexec/src/os/exec/exec.go:626 > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecuteContext({{0x140002a34b0?, 0x102341e88?, 0x14000482620?}}, {0x102362488?, 0x14000482620?}, {0x14000924840?, 0xc2331c84e20b9380?}, {0x10233ad80, 0x140007845c0}, {0x10233ad80, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:76 | cmd.Stderr = stderr | cmd.Stdout = stdout > return cmd.Run() | } | > github.com/deckhouse/virtualization/tests/e2e/executor.CMDExecutor.ExecContext({{0x140002a34b0?, 0x10384f728?, 0x103825fa0?}}, {0x102362488, 0x14000482620}, {0x14000924840, 0xb0}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/executor/executor.go:49 | stdout := new(Buffer) | stderr := new(Buffer) > err := e.ExecuteContext(ctx, command, stdout, stderr) | cmdResult := &CMDResult{ | stdOut: stdout, > github.com/deckhouse/virtualization/tests/e2e/kubectl.KubectlCMD.Get({{0x102366a38?, 0x1400000fc68?}, {0x10197ffba?, 0x140007a3c80?}}, {0x1019d72e2, 0x31}, {{0x0, 0x0, 0x0}, 0x0, ...}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/kubectl/kubectl.go:208 | ctx, cancel := context.WithTimeout(context.Background(), MediumTimeout) | defer cancel() > return k.ExecContext(ctx, cmd) | } | > github.com/deckhouse/virtualization/tests/e2e.SaveTestCaseResources(0x14000796b70, {0x14000719a28, 0x11}, {0x140006d9f50, 0x25}, {0x10197c370, 0x4}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:765 | errorFileName := fmt.Sprintf("%s/e2e_failed__%s__%s_error.txt", dumpPath, labels["testcase"], additional) | > cmdr := kubectl.Get( | "virtualization,cvi,vmc,intvirt,pod,volumesnapshot", | kc.GetOptions{ > github.com/deckhouse/virtualization/tests/e2e.SaveTestCaseDump(0x14000796b70, {0x101990e1e, 0x11}, {0x140006d9f50, 0x25}) /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:757 | } | > SaveTestCaseResources(labels, additional, namespace, tmpDir) | SavePodLogsAndDescriptions(labels, additional, namespace, tmpDir) | } > github.com/deckhouse/virtualization/tests/e2e.init.func12.2() /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:68 | AfterEach(func() { | if CurrentSpecReport().Failed() { > SaveTestCaseDump(testCaseLabel, CurrentSpecReport().LeafNodeText, ns) | } | }) github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?}) /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/node.go:475 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3() /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:894 github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 25 /Users/antont/go/pkg/mod/github.com/onsi/ginkgo/v2@v2.22.0/internal/suite.go:881 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/12/25 21:59:04.808 (3.237s) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/kubectl b/kubectl new file mode 100644 index 0000000000..ddd09d2ebe Binary files /dev/null and b/kubectl differ diff --git a/tests/e2e/artifacts/e2e/junit-20251001-102317.xml b/tests/e2e/artifacts/e2e/junit-20251001-102317.xml new file mode 100644 index 0000000000..d8a633a96f --- /dev/null +++ b/tests/e2e/artifacts/e2e/junit-20251001-102317.xml @@ -0,0 +1,2 @@ + + \ No newline at end of file diff --git a/tests/e2e/artifacts/e2e/junit-20251001-102343.xml b/tests/e2e/artifacts/e2e/junit-20251001-102343.xml new file mode 100644 index 0000000000..d8a633a96f --- /dev/null +++ b/tests/e2e/artifacts/e2e/junit-20251001-102343.xml @@ -0,0 +1,2 @@ + + \ No newline at end of file diff --git a/tests/e2e/artifacts/e2e/junit-20251001-102430.xml b/tests/e2e/artifacts/e2e/junit-20251001-102430.xml new file mode 100644 index 0000000000..a2610bd7be --- /dev/null +++ b/tests/e2e/artifacts/e2e/junit-20251001-102430.xml @@ -0,0 +1,764 @@ + + + + + + + + + + + + + + + + + + + + + + + + + > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/01/25 10:24:38.084 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/01/25 10:24:42.52 (4.436s) > Enter [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/01/25 10:24:42.52 < Exit [SynchronizedBeforeSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:137 @ 10/01/25 10:24:42.52 (0s) + + + > Enter [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/01/25 10:24:42.521 < Exit [BeforeAll] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:42 @ 10/01/25 10:24:44.3 (1.78s) > Enter [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/01/25 10:24:44.3 < Exit [BeforeEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:51 @ 10/01/25 10:24:50.967 (6.667s) > Enter [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/01/25 10:24:50.967 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:97 @ 10/01/25 10:24:50.968 STEP: Evacuate virtual machines - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:113 @ 10/01/25 10:26:31.314 STEP: Waiting for all VMOPs to be finished - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:116 @ 10/01/25 10:26:32.65 < Exit [It] Evacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:96 @ 10/01/25 10:26:47.768 (1m56.801s) > Enter [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/01/25 10:26:47.768 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 10:26:47.768 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 10:27:08.818 (21.051s) < Exit [AfterEach] VirtualMachineEvacuation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_evacuation_test.go:59 @ 10/01/25 10:27:08.818 (21.051s) + + + > Enter [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/01/25 10:27:08.818 < Exit [BeforeAll] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:57 @ 10/01/25 10:27:10.558 (1.74s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/01/25 10:27:10.558 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:73 @ 10/01/25 10:27:14.616 (4.058s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/01/25 10:27:14.616 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/01/25 10:27:14.616 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/01/25 10:27:14.617 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:97 @ 10/01/25 10:27:14.617 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:96 @ 10/01/25 10:27:29.112 (14.496s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/01/25 10:27:29.112 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/01/25 10:27:29.112 (0s) + + + > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/01/25 10:27:29.113 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:108 @ 10/01/25 10:27:29.113 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:107 @ 10/01/25 10:28:47.668 (1m18.555s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/01/25 10:28:47.668 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/01/25 10:28:47.668 (0s) + + + > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:118 @ 10/01/25 10:28:47.668 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:119 @ 10/01/25 10:28:47.668 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:118 @ 10/01/25 10:30:01.74 (1m14.072s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/01/25 10:30:01.74 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/01/25 10:30:01.74 (0s) + + + > Enter [It] status should be in Running phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:129 @ 10/01/25 10:30:01.74 < Exit [It] status should be in Running phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:129 @ 10/01/25 10:30:05.393 (3.653s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/01/25 10:30:05.393 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/01/25 10:30:05.393 (0s) + + + > Enter [It] gets VMs and SVCs objects - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:142 @ 10/01/25 10:30:05.393 < Exit [It] gets VMs and SVCs objects - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:142 @ 10/01/25 10:30:09.361 (3.968s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/01/25 10:30:09.362 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/01/25 10:30:09.362 (0s) + + + [FAILED] Timed out after 90.001s. The function passed to Eventually returned the following error: <*errors.errorString | 0x14000a01b60>: cmd: d8 v ssh head-45bdb83d-vm-connectivity-a -c 'hostname' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-connectivity --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed stderr: bash: d8: command not found { s: "cmd: d8 v ssh head-45bdb83d-vm-connectivity-a -c 'hostname' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-connectivity --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed\nstderr: bash: d8: command not found\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:170 @ 10/01/25 10:31:39.363 + > Enter [It] check ssh connection via `d8 v` to VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:166 @ 10/01/25 10:30:09.362 STEP: VirtualMachine "head-45bdb83d-vm-connectivity-a" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:169 @ 10/01/25 10:30:09.362 [FAILED] Timed out after 90.001s. The function passed to Eventually returned the following error: <*errors.errorString | 0x14000a01b60>: cmd: d8 v ssh head-45bdb83d-vm-connectivity-a -c 'hostname' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-connectivity --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed stderr: bash: d8: command not found { s: "cmd: d8 v ssh head-45bdb83d-vm-connectivity-a -c 'hostname' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-connectivity --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed\nstderr: bash: d8: command not found\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:170 @ 10/01/25 10:31:39.363 < Exit [It] check ssh connection via `d8 v` to VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:166 @ 10/01/25 10:31:39.363 (1m30.002s) > Enter [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/01/25 10:31:39.363 < Exit [AfterEach] VirtualMachineConnectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:66 @ 10/01/25 10:31:44.511 (5.147s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:174 @ 10/01/25 10:31:44.511 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:179 @ 10/01/25 10:31:44.511 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:187 @ 10/01/25 10:31:44.511 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:198 @ 10/01/25 10:31:44.511 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:209 @ 10/01/25 10:31:44.511 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:222 @ 10/01/25 10:31:44.511 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:228 @ 10/01/25 10:31:44.511 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:240 @ 10/01/25 10:31:44.511 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:250 @ 10/01/25 10:31:44.511 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_connectivity_test.go:258 @ 10/01/25 10:31:44.511 + + + > Enter [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/01/25 10:31:44.511 < Exit [BeforeAll] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:55 @ 10/01/25 10:31:46.259 (1.748s) > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/01/25 10:31:46.259 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/01/25 10:31:46.259 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/01/25 10:31:46.259 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:80 @ 10/01/25 10:31:53.44 (7.18s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/01/25 10:31:53.44 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/01/25 10:31:53.44 (0s) + + + > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/01/25 10:31:53.44 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/01/25 10:31:53.44 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/01/25 10:31:53.44 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/01/25 10:31:53.44 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:112 @ 10/01/25 10:33:13.859 (1m20.42s) STEP: `VirtualMachineBlockDeviceAttachment` should be attached - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:119 @ 10/01/25 10:33:13.859 END STEP: `VirtualMachineBlockDeviceAttachment` should be attached - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:119 @ 10/01/25 10:33:15.925 (2.065s) < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:111 @ 10/01/25 10:33:15.925 (1m22.485s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/01/25 10:33:15.925 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/01/25 10:33:15.925 (0s) + + + > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/01/25 10:33:15.925 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/01/25 10:33:15.925 (0s) > Enter [It] add additional interface to virtual machines - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:131 @ 10/01/25 10:33:15.925 STEP: patch virtual machines for add additional network interface - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:137 @ 10/01/25 10:33:16.049 END STEP: patch virtual machines for add additional network interface - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:137 @ 10/01/25 10:33:18.943 (2.893s) STEP: `VirtualMachine` agent should be ready after patching - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:155 @ 10/01/25 10:33:18.943 END STEP: `VirtualMachine` agent should be ready after patching - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:155 @ 10/01/25 10:34:00.785 (41.838s) STEP: remembering the .status.networks of each VM after patching - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:163 @ 10/01/25 10:34:00.785 END STEP: remembering the .status.networks of each VM after patching - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:163 @ 10/01/25 10:34:01.757 (971ms) < Exit [It] add additional interface to virtual machines - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:131 @ 10/01/25 10:34:01.757 (45.827s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/01/25 10:34:01.757 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/01/25 10:34:01.757 (0s) + + + > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/01/25 10:34:01.757 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/01/25 10:34:01.757 (0s) > Enter [It] restore the `VirtualMachines` with `forced` mode - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:177 @ 10/01/25 10:34:01.757 STEP: Getting `VirtualMachines` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:181 @ 10/01/25 10:34:01.757 END STEP: Getting `VirtualMachines` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:181 @ 10/01/25 10:34:02.302 (544ms) STEP: Creating `VirtualMachineSnapshots` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:189 @ 10/01/25 10:34:02.302 END STEP: Creating `VirtualMachineSnapshots` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:189 @ 10/01/25 10:34:13.016 (10.708s) STEP: Attaching `VirtualDisk` after `VirtualMachine` snapshotting - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:209 @ 10/01/25 10:34:13.016 END STEP: Attaching `VirtualDisk` after `VirtualMachine` snapshotting - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:209 @ 10/01/25 10:34:39.783 (26.762s) STEP: Creating `VirtualMachineRestores` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:231 @ 10/01/25 10:34:39.783 END STEP: Creating `VirtualMachineRestores` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:231 @ 10/01/25 10:39:08.565 (4m28.781s) STEP: Checking the result of restoration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:292 @ 10/01/25 10:39:08.566 END STEP: Checking the result of restoration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:292 @ 10/01/25 10:39:19.636 (11.071s) < Exit [It] restore the `VirtualMachines` with `forced` mode - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:177 @ 10/01/25 10:39:19.636 (5m17.866s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/01/25 10:39:19.636 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/01/25 10:39:19.636 (0s) + + + > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/01/25 10:39:19.637 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/01/25 10:39:19.637 (0s) > Enter [It] check the .status.networks of each VM after restore - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:341 @ 10/01/25 10:39:19.637 < Exit [It] check the .status.networks of each VM after restore - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:341 @ 10/01/25 10:39:24.136 (4.5s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/01/25 10:39:24.136 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/01/25 10:39:24.136 (0s) + + + > Enter [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/01/25 10:39:24.137 < Exit [BeforeEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:64 @ 10/01/25 10:39:24.137 (0s) > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:367 @ 10/01/25 10:39:24.137 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 10:39:24.137 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 10:39:51.28 (27.143s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:367 @ 10/01/25 10:39:51.28 (27.143s) > Enter [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/01/25 10:39:51.28 < Exit [AfterEach] VirtualMachineRestoreForce - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_force_test.go:71 @ 10/01/25 10:39:51.28 (0s) + + + > Enter [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/01/25 10:39:51.28 < Exit [BeforeAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:36 @ 10/01/25 10:39:51.282 (2ms) > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/01/25 10:39:51.282 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/01/25 10:39:51.282 (0s) > Enter [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/01/25 10:39:51.282 < Exit [It] creates project - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:61 @ 10/01/25 10:39:52.892 (1.61s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/01/25 10:39:52.892 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/01/25 10:39:52.892 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/01/25 10:39:52.892 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/01/25 10:39:52.892 (0s) > Enter [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/01/25 10:39:52.892 STEP: Project should be deployed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:73 @ 10/01/25 10:39:52.892 < Exit [It] checks project readiness - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:72 @ 10/01/25 10:39:55.138 (2.246s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/01/25 10:39:55.138 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/01/25 10:39:55.138 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/01/25 10:39:55.139 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/01/25 10:39:55.139 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/01/25 10:39:55.139 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:83 @ 10/01/25 10:39:58.13 (2.991s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/01/25 10:39:58.13 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/01/25 10:39:58.13 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/01/25 10:39:58.13 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/01/25 10:39:58.13 (0s) > Enter [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/01/25 10:39:58.13 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/01/25 10:39:58.13 < Exit [It] When virtual images are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:101 @ 10/01/25 10:40:04.438 (6.308s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/01/25 10:40:04.438 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/01/25 10:40:04.438 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/01/25 10:40:04.438 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/01/25 10:40:04.438 (0s) > Enter [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/01/25 10:40:04.438 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/01/25 10:40:04.438 < Exit [It] When virtual disks are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:102 @ 10/01/25 10:40:42.036 (37.598s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/01/25 10:40:42.036 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/01/25 10:40:42.036 (0s) + + + > Enter [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/01/25 10:40:42.036 < Exit [BeforeEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:43 @ 10/01/25 10:40:42.036 (0s) > Enter [It] When virtual machines are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/01/25 10:40:42.036 STEP: VMs should be in Running phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:94 @ 10/01/25 10:40:42.037 < Exit [It] When virtual machines are applied - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:103 @ 10/01/25 10:41:11.095 (29.058s) > Enter [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/01/25 10:41:11.095 < Exit [AfterEach] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:54 @ 10/01/25 10:41:11.095 (0s) > Enter [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/01/25 10:41:11.095 STEP: Delete manifests - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:50 @ 10/01/25 10:41:11.095 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 10:41:11.095 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 10:42:15.243 (1m4.148s) < Exit [AfterAll] ImporterNetworkPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/importer_network_policy_test.go:49 @ 10/01/25 10:42:15.243 (1m4.148s) + + + > Enter [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/01/25 10:42:15.243 < Exit [BeforeAll] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:49 @ 10/01/25 10:42:17.02 (1.777s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/01/25 10:42:17.02 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:65 @ 10/01/25 10:42:22.141 (5.121s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/01/25 10:42:22.141 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/01/25 10:42:22.141 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/01/25 10:42:22.142 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:89 @ 10/01/25 10:42:22.142 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:88 @ 10/01/25 10:42:28.533 (6.391s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/01/25 10:42:28.533 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/01/25 10:42:28.533 (0s) + + + > Enter [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/01/25 10:42:28.533 < Exit [It] should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:99 @ 10/01/25 10:43:09.306 (40.773s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/01/25 10:43:09.306 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/01/25 10:43:09.306 (0s) + + + > Enter [It] should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:109 @ 10/01/25 10:43:09.306 < Exit [It] should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:109 @ 10/01/25 10:44:02.366 (53.06s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/01/25 10:44:02.366 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/01/25 10:44:02.366 (0s) + + + > Enter [It] changes the number of processor cores - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:123 @ 10/01/25 10:44:02.366 STEP: Checking the number of processor cores before changing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:303 @ 10/01/25 10:44:04.381 STEP: Patching virtual machine specification - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:296 @ 10/01/25 10:44:05.396 < Exit [It] changes the number of processor cores - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:123 @ 10/01/25 10:44:06.512 (4.146s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/01/25 10:44:06.512 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/01/25 10:44:06.512 (0s) + + + > Enter [It] checks the number of processor cores in specification - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:147 @ 10/01/25 10:44:06.513 STEP: Checking the number of processor cores after changing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:303 @ 10/01/25 10:44:07.773 < Exit [It] checks the number of processor cores in specification - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:147 @ 10/01/25 10:44:09.021 (2.508s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/01/25 10:44:09.021 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/01/25 10:44:09.021 (0s) + + + [FAILED] Timed out after 90.001s. Unexpected error: <*errors.errorString | 0x14001188800>: cmd: d8 v ssh head-45bdb83d-vm-manual-conf -c 'sudo nohup reboot -f > /dev/null 2>&1 &' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-vm-configuration --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed stderr: bash: d8: command not found { s: "cmd: d8 v ssh head-45bdb83d-vm-manual-conf -c 'sudo nohup reboot -f > /dev/null 2>&1 &' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-vm-configuration --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed\nstderr: bash: d8: command not found\n", } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:172 @ 10/01/25 10:45:40.275 + > Enter [It] should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:161 @ 10/01/25 10:44:09.021 [FAILED] Timed out after 90.001s. Unexpected error: <*errors.errorString | 0x14001188800>: cmd: d8 v ssh head-45bdb83d-vm-manual-conf -c 'sudo nohup reboot -f > /dev/null 2>&1 &' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-vm-configuration --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed stderr: bash: d8: command not found { s: "cmd: d8 v ssh head-45bdb83d-vm-manual-conf -c 'sudo nohup reboot -f > /dev/null 2>&1 &' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-vm-configuration --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed\nstderr: bash: d8: command not found\n", } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:172 @ 10/01/25 10:45:40.275 < Exit [It] should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:161 @ 10/01/25 10:45:40.276 (1m31.255s) > Enter [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/01/25 10:45:40.276 < Exit [AfterEach] VirtualMachineConfiguration 1 - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:58 @ 10/01/25 10:45:46.04 (5.765s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:183 @ 10/01/25 10:45:46.041 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:202 @ 10/01/25 10:45:46.041 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:226 @ 10/01/25 10:45:46.041 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:240 @ 10/01/25 10:45:46.041 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:250 @ 10/01/25 10:45:46.041 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_configuration_test.go:265 @ 10/01/25 10:45:46.041 + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.041 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/01/25 10:45:46.041 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.041 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.041 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/01/25 10:45:46.041 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.041 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.041 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/01/25 10:45:46.041 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.041 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.042 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/01/25 10:45:46.042 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.042 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.042 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/01/25 10:45:46.042 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.042 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.042 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/01/25 10:45:46.042 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.042 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.042 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/01/25 10:45:46.042 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.042 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.042 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/01/25 10:45:46.042 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.042 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.042 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/01/25 10:45:46.042 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.042 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.042 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/01/25 10:45:46.042 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.042 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.042 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/01/25 10:45:46.042 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.042 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.043 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:55 @ 10/01/25 10:45:46.043 < Exit [BeforeEach] [sig-storage] Volume migration with local disks - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_local_disks.go:53 @ 10/01/25 10:45:46.043 (0s) + + + > Enter [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/01/25 10:45:46.043 < Exit [BeforeAll] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:52 @ 10/01/25 10:45:47.797 (1.754s) > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:45:47.797 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:45:47.797 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/01/25 10:45:47.797 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:80 @ 10/01/25 10:45:55.135 (7.337s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:45:55.135 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:45:55.135 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:45:55.135 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:45:55.135 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/01/25 10:45:55.135 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:91 @ 10/01/25 10:45:55.135 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:90 @ 10/01/25 10:45:59.375 (4.24s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:45:59.375 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:45:59.375 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:45:59.375 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:45:59.375 (0s) > Enter [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/01/25 10:45:59.375 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:102 @ 10/01/25 10:45:59.375 < Exit [It] checks VDs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:101 @ 10/01/25 10:46:47.028 (47.653s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:46:47.028 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:46:47.028 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:46:47.028 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:46:47.028 (0s) > Enter [It] checks VDs phases with map[vm:existing-vmclass] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/01/25 10:46:47.028 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:116 @ 10/01/25 10:46:47.028 < Exit [It] checks VDs phases with map[vm:existing-vmclass] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:115 @ 10/01/25 10:46:49.322 (2.294s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:46:49.322 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:46:49.322 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:46:49.322 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:46:49.322 (0s) > Enter [It] checks VMs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:126 @ 10/01/25 10:46:49.322 STEP: VMs should be in Pending phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:127 @ 10/01/25 10:46:49.322 < Exit [It] checks VMs phases with map[vm:not-existing-vmclass-with-changing] and map[vm:not-existing-vmclass-with-creating] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:126 @ 10/01/25 10:46:53.908 (4.586s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:46:53.908 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:46:53.908 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:46:53.908 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:46:53.909 (0s) > Enter [It] checks VMs phases with map[vm:existing-vmclass] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:140 @ 10/01/25 10:46:53.909 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:141 @ 10/01/25 10:46:53.909 < Exit [It] checks VMs phases with map[vm:existing-vmclass] label - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:140 @ 10/01/25 10:47:41.367 (47.458s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:47:41.367 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:47:41.367 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:47:41.367 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:47:41.367 (0s) > Enter [It] checks condition status before changing 'virtulaMachineCLass` field with existing class - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:152 @ 10/01/25 10:47:41.367 STEP: VirtualMachineClassReady status should be 'False' before changing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:153 @ 10/01/25 10:47:41.367 < Exit [It] checks condition status before changing 'virtulaMachineCLass` field with existing class - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:152 @ 10/01/25 10:47:42.663 (1.296s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:47:42.663 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:47:42.663 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:47:42.663 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:47:42.663 (0s) > Enter [It] changes VMClassName in VM specification with existing VMClass - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:157 @ 10/01/25 10:47:42.663 < Exit [It] changes VMClassName in VM specification with existing VMClass - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:157 @ 10/01/25 10:47:44.073 (1.41s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:47:44.073 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:47:44.073 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:47:44.073 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:47:44.073 (0s) > Enter [It] checks VM phase and condition status after changing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:163 @ 10/01/25 10:47:44.073 STEP: VM should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:164 @ 10/01/25 10:47:44.073 STEP: VirtualMachineClassReady status should be 'True' after changing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:170 @ 10/01/25 10:48:27.246 < Exit [It] checks VM phase and condition status after changing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:163 @ 10/01/25 10:48:28.165 (44.092s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:48:28.165 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:48:28.165 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:48:28.165 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:48:28.165 (0s) > Enter [It] checks condition status before creating `VirtualMachineClass` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:176 @ 10/01/25 10:48:28.165 STEP: VirtualMachineClassReady status should be 'False' before creating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:177 @ 10/01/25 10:48:28.165 < Exit [It] checks condition status before creating `VirtualMachineClass` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:176 @ 10/01/25 10:48:29.25 (1.085s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:48:29.25 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:48:29.25 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:48:29.25 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:48:29.25 (0s) > Enter [It] changes VMClassName in VM specification with not existing VMClass which have correct prefix for creating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:181 @ 10/01/25 10:48:29.25 < Exit [It] changes VMClassName in VM specification with not existing VMClass which have correct prefix for creating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:181 @ 10/01/25 10:48:30.32 (1.07s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:48:30.32 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:48:30.32 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:48:30.321 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:48:30.321 (0s) > Enter [It] creates new `VirtualMachineClass` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:187 @ 10/01/25 10:48:30.321 < Exit [It] creates new `VirtualMachineClass` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:187 @ 10/01/25 10:48:32.898 (2.578s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:48:32.898 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:48:32.898 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:48:32.899 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:48:32.899 (0s) > Enter [It] checks VM phase and condition after creating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:202 @ 10/01/25 10:48:32.899 STEP: VM should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:203 @ 10/01/25 10:48:32.899 STEP: VirtualMachineClassReady status should be 'True' after creating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:209 @ 10/01/25 10:49:17.424 < Exit [It] checks VM phase and condition after creating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:202 @ 10/01/25 10:49:18.388 (45.489s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:49:18.388 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:49:18.388 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:49:18.389 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:49:18.389 (0s) > Enter [It] checks sizing policy match - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:216 @ 10/01/25 10:49:18.389 STEP: Check virtual machine: head-45bdb83d-vm-existing-vmclass - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:230 @ 10/01/25 10:49:20.372 STEP: Check virtual machine: head-45bdb83d-vm-not-existing-vmclass-with-changing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:230 @ 10/01/25 10:49:21.479 STEP: Check virtual machine: head-45bdb83d-vm-not-existing-vmclass-with-creating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:230 @ 10/01/25 10:49:22.388 < Exit [It] checks sizing policy match - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:216 @ 10/01/25 10:49:23.479 (5.091s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:49:23.479 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:49:23.479 (0s) + + + > Enter [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:49:23.48 < Exit [BeforeEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:67 @ 10/01/25 10:49:23.48 (0s) > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:240 @ 10/01/25 10:49:23.48 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 10:49:23.48 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 10:49:50.569 (27.086s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:240 @ 10/01/25 10:49:50.569 (27.086s) > Enter [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:49:50.569 < Exit [AfterEach] SizingPolicy - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/sizing_policy_test.go:73 @ 10/01/25 10:49:50.569 (0s) + + + > Enter [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/01/25 10:49:50.569 < Exit [BeforeAll] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:35 @ 10/01/25 10:49:52.328 (1.757s) > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/01/25 10:49:52.328 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/01/25 10:49:52.328 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/01/25 10:49:52.328 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:57 @ 10/01/25 10:49:54.407 (2.078s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/01/25 10:49:54.407 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/01/25 10:49:54.407 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/01/25 10:49:54.407 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/01/25 10:49:54.407 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/01/25 10:49:54.407 STEP: VDs should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:68 @ 10/01/25 10:49:54.407 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:67 @ 10/01/25 10:50:09.051 (14.64s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/01/25 10:50:09.051 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/01/25 10:50:09.051 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/01/25 10:50:09.052 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/01/25 10:50:09.052 (0s) > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/01/25 10:50:09.052 STEP: VM should be in Running phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:79 @ 10/01/25 10:50:09.052 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:78 @ 10/01/25 10:51:05.079 (56.025s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/01/25 10:51:05.079 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/01/25 10:51:05.079 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/01/25 10:51:05.08 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/01/25 10:51:05.08 (0s) > Enter [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/01/25 10:51:05.08 < Exit [It] has qemu version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:97 @ 10/01/25 10:51:05.08 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/01/25 10:51:05.08 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/01/25 10:51:05.08 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/01/25 10:51:05.08 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/01/25 10:51:05.08 (0s) > Enter [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/01/25 10:51:05.08 < Exit [It] has libvirt version in the status - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:103 @ 10/01/25 10:51:05.08 (0s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/01/25 10:51:05.08 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/01/25 10:51:05.08 (0s) + + + > Enter [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/01/25 10:51:05.08 < Exit [BeforeEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:44 @ 10/01/25 10:51:05.08 (0s) > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/01/25 10:51:05.08 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 10:51:05.08 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 10:52:05.405 (1m0.324s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:114 @ 10/01/25 10:52:05.405 (1m0.324s) > Enter [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/01/25 10:52:05.405 < Exit [AfterEach] VirtualMachineVersions - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_version_test.go:50 @ 10/01/25 10:52:05.405 (0s) + + + + > Enter [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/01/25 10:52:05.405 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:41 @ 10/01/25 10:52:05.406 < Exit [BeforeAll] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:39 @ 10/01/25 10:52:05.406 (1ms) > Enter [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/01/25 10:52:05.406 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 10:52:05.406 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 10:52:09.888 (4.482s) < Exit [AfterEach] VirtualMachineCancelMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_cancel_test.go:59 @ 10/01/25 10:52:09.888 (4.482s) + + + > Enter [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/01/25 10:52:09.888 < Exit [BeforeAll] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:52 @ 10/01/25 10:52:11.677 (1.789s) > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/01/25 10:52:11.677 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/01/25 10:52:11.677 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/01/25 10:52:11.677 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:74 @ 10/01/25 10:52:16.218 (4.541s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/01/25 10:52:16.218 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/01/25 10:52:16.218 (0s) + + + > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/01/25 10:52:16.218 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/01/25 10:52:16.218 (0s) > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/01/25 10:52:16.218 STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/01/25 10:52:16.218 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:106 @ 10/01/25 10:53:32.416 (1m16.198s) < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:105 @ 10/01/25 10:53:32.416 (1m16.198s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/01/25 10:53:32.416 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/01/25 10:53:32.417 (0s) + + + > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/01/25 10:53:32.417 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/01/25 10:53:32.417 (0s) > Enter [It] add additional interface to virtual machines - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:115 @ 10/01/25 10:53:32.417 STEP: patch virtual machine for add additional network interface - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:120 @ 10/01/25 10:53:33.123 END STEP: patch virtual machine for add additional network interface - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:120 @ 10/01/25 10:53:35.857 (2.733s) STEP: `VirtualMachine` agent should be ready after patching - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:137 @ 10/01/25 10:53:35.857 END STEP: `VirtualMachine` agent should be ready after patching - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:137 @ 10/01/25 10:54:15.879 (40.022s) STEP: remembering the .status.networks of each VM after patching - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:144 @ 10/01/25 10:54:15.879 END STEP: remembering the .status.networks of each VM after patching - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:144 @ 10/01/25 10:54:16.475 (595ms) < Exit [It] add additional interface to virtual machines - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:115 @ 10/01/25 10:54:16.475 (44.058s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/01/25 10:54:16.475 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/01/25 10:54:16.475 (0s) + + + > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/01/25 10:54:16.475 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/01/25 10:54:16.475 (0s) > Enter [It] restore the `VirtualMachines` with `Safe` mode - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:158 @ 10/01/25 10:54:16.475 STEP: Getting `VirtualMachines` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:162 @ 10/01/25 10:54:16.475 END STEP: Getting `VirtualMachines` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:162 @ 10/01/25 10:54:17.375 (900ms) STEP: Creating `VirtualMachineSnapshots` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:170 @ 10/01/25 10:54:17.375 END STEP: Creating `VirtualMachineSnapshots` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:170 @ 10/01/25 10:54:25.043 (7.668s) STEP: Attaching `VirtualDisk` after `VirtualMachine` snapshotting - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:190 @ 10/01/25 10:54:25.043 END STEP: Attaching `VirtualDisk` after `VirtualMachine` snapshotting - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:190 @ 10/01/25 10:54:35.528 (10.485s) STEP: Deleting `VirtualMachines` and their resources for `Safe` restoring - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:212 @ 10/01/25 10:54:35.528 END STEP: Deleting `VirtualMachines` and their resources for `Safe` restoring - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:212 @ 10/01/25 10:54:58.709 (23.181s) STEP: Creating `VirtualMachineRestores` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:275 @ 10/01/25 10:54:58.709 END STEP: Creating `VirtualMachineRestores` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:275 @ 10/01/25 10:56:18.185 (1m19.475s) STEP: Checking the result of restoration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:300 @ 10/01/25 10:56:18.185 END STEP: Checking the result of restoration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:300 @ 10/01/25 10:56:23.648 (5.464s) < Exit [It] restore the `VirtualMachines` with `Safe` mode - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:158 @ 10/01/25 10:56:23.648 (2m7.173s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/01/25 10:56:23.648 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/01/25 10:56:23.648 (0s) + + + > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/01/25 10:56:23.649 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/01/25 10:56:23.649 (0s) > Enter [It] check the .status.networks of each VM after restore - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:348 @ 10/01/25 10:56:23.649 < Exit [It] check the .status.networks of each VM after restore - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:348 @ 10/01/25 10:56:26.641 (2.992s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/01/25 10:56:26.641 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/01/25 10:56:26.641 (0s) + + + > Enter [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/01/25 10:56:26.641 < Exit [BeforeEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:61 @ 10/01/25 10:56:26.641 (0s) > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:374 @ 10/01/25 10:56:26.641 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 10:56:26.641 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 10:56:34.654 (8.013s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:374 @ 10/01/25 10:56:34.654 (8.013s) > Enter [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/01/25 10:56:34.654 < Exit [AfterEach] VirtualMachineRestoreSafe - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_restore_safe_test.go:65 @ 10/01/25 10:56:34.654 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.654 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/01/25 10:56:34.655 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.655 (1ms) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.655 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/01/25 10:56:34.655 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.655 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.655 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/01/25 10:56:34.655 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.655 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.655 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/01/25 10:56:34.655 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.655 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.655 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/01/25 10:56:34.656 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.656 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.656 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/01/25 10:56:34.656 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.656 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.656 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/01/25 10:56:34.656 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.656 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.656 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/01/25 10:56:34.657 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.657 (0s) + + + + > Enter [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.657 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeEach] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:52 @ 10/01/25 10:56:34.657 < Exit [BeforeEach] [sig-storage] Volume migration when storage class changed - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/storage/volume_migration_storage_class_changed.go:50 @ 10/01/25 10:56:34.657 (0s) + + + > Enter [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/01/25 10:56:34.657 < Exit [BeforeAll] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:52 @ 10/01/25 10:56:36.38 (1.723s) > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/01/25 10:56:36.38 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/01/25 10:56:36.38 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/01/25 10:56:36.38 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:77 @ 10/01/25 10:56:39.75 (3.37s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/01/25 10:56:39.751 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/01/25 10:56:39.751 (0s) + + + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/01/25 10:56:39.751 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/01/25 10:56:39.751 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/01/25 10:56:39.751 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:88 @ 10/01/25 10:56:39.751 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:87 @ 10/01/25 10:56:45.415 (5.664s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/01/25 10:56:45.415 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/01/25 10:56:45.415 (0s) + + + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/01/25 10:56:45.415 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/01/25 10:56:45.415 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/01/25 10:56:45.415 STEP: VDs with consumers should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:99 @ 10/01/25 10:56:45.415 STEP: VDs without consumers should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:106 @ 10/01/25 10:56:55.759 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:98 @ 10/01/25 10:56:58.033 (12.618s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/01/25 10:56:58.033 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/01/25 10:56:58.033 (0s) + + + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/01/25 10:56:58.034 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/01/25 10:56:58.034 (0s) > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/01/25 10:56:58.034 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:117 @ 10/01/25 10:56:58.034 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:116 @ 10/01/25 10:58:17.806 (1m19.772s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/01/25 10:58:17.806 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/01/25 10:58:17.807 (0s) + + + [FAILED] Timed out after 90.001s. virtualMachine: head-45bdb83d-vm-automatic-with-hotplug-standalone Unexpected error: <*errors.errorString | 0x14001295f30>: cmd: d8 v ssh head-45bdb83d-vm-automatic-with-hotplug-standalone -c 'lsblk --nodeps --json' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-vm-disk-attachment --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed stderr: bash: d8: command not found { s: "cmd: d8 v ssh head-45bdb83d-vm-automatic-with-hotplug-standalone -c 'lsblk --nodeps --json' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-vm-disk-attachment --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed\nstderr: bash: d8: command not found\n", } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:131 @ 10/01/25 10:59:47.809 + > Enter [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/01/25 10:58:17.807 < Exit [BeforeEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:64 @ 10/01/25 10:58:17.807 (0s) > Enter [It] get disk count before attachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/01/25 10:58:17.807 [FAILED] Timed out after 90.001s. virtualMachine: head-45bdb83d-vm-automatic-with-hotplug-standalone Unexpected error: <*errors.errorString | 0x14001295f30>: cmd: d8 v ssh head-45bdb83d-vm-automatic-with-hotplug-standalone -c 'lsblk --nodeps --json' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-vm-disk-attachment --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed stderr: bash: d8: command not found { s: "cmd: d8 v ssh head-45bdb83d-vm-automatic-with-hotplug-standalone -c 'lsblk --nodeps --json' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-vm-disk-attachment --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed\nstderr: bash: d8: command not found\n", } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:131 @ 10/01/25 10:59:47.809 < Exit [It] get disk count before attachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:128 @ 10/01/25 10:59:47.809 (1m30.002s) > Enter [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/01/25 10:59:47.809 < Exit [AfterEach] VirtualDiskAttachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:70 @ 10/01/25 10:59:53.752 (5.943s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:133 @ 10/01/25 10:59:53.752 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:136 @ 10/01/25 10:59:53.752 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:150 @ 10/01/25 10:59:53.752 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:166 @ 10/01/25 10:59:53.752 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:171 @ 10/01/25 10:59:53.752 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:179 @ 10/01/25 10:59:53.752 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:187 @ 10/01/25 10:59:53.752 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_attachment_test.go:201 @ 10/01/25 10:59:53.752 + + + > Enter [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/01/25 10:59:53.752 < Exit [BeforeAll] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:54 @ 10/01/25 10:59:56.431 (2.679s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/01/25 10:59:56.431 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:77 @ 10/01/25 11:00:00.521 (4.089s) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/01/25 11:00:00.521 < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/01/25 11:00:00.521 (0s) + + + > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:112 @ 10/01/25 11:00:00.521 STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:113 @ 10/01/25 11:00:00.521 END STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:113 @ 10/01/25 11:00:14.348 (13.827s) STEP: `ClusterVirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:120 @ 10/01/25 11:00:14.348 END STEP: `ClusterVirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:120 @ 10/01/25 11:00:59.154 (44.806s) STEP: `VirtualDisk` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:127 @ 10/01/25 11:00:59.155 END STEP: `VirtualDisk` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:127 @ 10/01/25 11:01:01.934 (2.779s) STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:134 @ 10/01/25 11:01:01.934 END STEP: `VirtualMachine` agent should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:134 @ 10/01/25 11:01:19.666 (17.732s) < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:112 @ 10/01/25 11:01:19.666 (1m19.145s) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/01/25 11:01:19.666 < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/01/25 11:01:19.666 (0s) + + + > Enter [It] retrieves the test objects - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:147 @ 10/01/25 11:01:19.666 STEP: `VirtualMachine` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:148 @ 10/01/25 11:01:19.666 END STEP: `VirtualMachine` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:148 @ 10/01/25 11:01:20.288 (621ms) STEP: `VirtualImages` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:158 @ 10/01/25 11:01:20.288 END STEP: `VirtualImages` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:158 @ 10/01/25 11:01:21.184 (896ms) STEP: `ClusterVirtualImages` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:173 @ 10/01/25 11:01:21.184 END STEP: `ClusterVirtualImages` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:173 @ 10/01/25 11:01:21.776 (591ms) < Exit [It] retrieves the test objects - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:147 @ 10/01/25 11:01:21.776 (2.109s) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/01/25 11:01:21.776 < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/01/25 11:01:21.776 (0s) + + + [FAILED] Timed out after 90.000s. virtualMachine: head-45bdb83d-vm-image-hotplug Unexpected error: <*errors.errorString | 0x14001295940>: cmd: d8 v ssh head-45bdb83d-vm-image-hotplug -c 'lsblk --nodeps --json' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-image-hotplug --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed stderr: bash: d8: command not found { s: "cmd: d8 v ssh head-45bdb83d-vm-image-hotplug -c 'lsblk --nodeps --json' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-image-hotplug --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed\nstderr: bash: d8: command not found\n", } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:193 @ 10/01/25 11:02:51.777 + > Enter [It] retrieves the disk count before the images attachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:190 @ 10/01/25 11:01:21.776 [FAILED] Timed out after 90.000s. virtualMachine: head-45bdb83d-vm-image-hotplug Unexpected error: <*errors.errorString | 0x14001295940>: cmd: d8 v ssh head-45bdb83d-vm-image-hotplug -c 'lsblk --nodeps --json' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-image-hotplug --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed stderr: bash: d8: command not found { s: "cmd: d8 v ssh head-45bdb83d-vm-image-hotplug -c 'lsblk --nodeps --json' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-image-hotplug --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed\nstderr: bash: d8: command not found\n", } occurred In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:193 @ 10/01/25 11:02:51.777 < Exit [It] retrieves the disk count before the images attachment - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:190 @ 10/01/25 11:02:51.777 (1m30.001s) > Enter [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/01/25 11:02:51.777 < Exit [AfterEach] ImageHotplug - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:70 @ 10/01/25 11:02:58.488 (6.71s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:196 @ 10/01/25 11:02:58.488 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:204 @ 10/01/25 11:02:58.488 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:224 @ 10/01/25 11:02:58.488 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:236 @ 10/01/25 11:02:58.488 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:258 @ 10/01/25 11:02:58.488 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:284 @ 10/01/25 11:02:58.488 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:294 @ 10/01/25 11:02:58.488 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/image_hotplug_test.go:309 @ 10/01/25 11:02:58.488 + + + > Enter [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:54 @ 10/01/25 11:02:58.489 < Exit [BeforeAll] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:54 @ 10/01/25 11:03:00.248 (1.76s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:85 @ 10/01/25 11:03:00.248 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:85 @ 10/01/25 11:03:05.993 (5.744s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:03:05.993 < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:03:05.993 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:95 @ 10/01/25 11:03:05.993 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:96 @ 10/01/25 11:03:05.993 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:95 @ 10/01/25 11:03:17.119 (11.126s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:03:17.119 < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:03:17.119 (0s) + + + > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:106 @ 10/01/25 11:03:17.119 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:107 @ 10/01/25 11:03:17.119 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:106 @ 10/01/25 11:04:09.07 (51.951s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:04:09.07 < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:04:09.07 (0s) + + + > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:117 @ 10/01/25 11:04:09.07 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:118 @ 10/01/25 11:04:09.07 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:117 @ 10/01/25 11:05:29.315 (1m20.244s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:05:29.315 < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:05:29.315 (0s) + + + > Enter [It] checks VMBDAs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:128 @ 10/01/25 11:05:29.315 STEP: VMBDAs should be in Attached phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:129 @ 10/01/25 11:05:29.315 < Exit [It] checks VMBDAs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:128 @ 10/01/25 11:05:31.561 (2.246s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:05:31.561 < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:05:31.561 (0s) + + + > Enter [It] creates VDs snapshots with `requiredConsistency` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:139 @ 10/01/25 11:05:31.561 STEP: Create snapshot for "head-45bdb83d-vd-ubuntu-http" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:150 @ 10/01/25 11:05:32.717 < Exit [It] creates VDs snapshots with `requiredConsistency` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:139 @ 10/01/25 11:05:33.882 (2.321s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:05:33.882 < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:05:33.882 (0s) + + + > Enter [It] checks snapshots of unattached VDs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:159 @ 10/01/25 11:05:33.883 STEP: Snapshots should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:160 @ 10/01/25 11:05:33.883 < Exit [It] checks snapshots of unattached VDs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:159 @ 10/01/25 11:05:36.108 (2.225s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:05:36.108 < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:05:36.108 (0s) + + + > Enter [It] creates snapshots with `requiredConsistency` of attached VDs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:180 @ 10/01/25 11:05:36.108 STEP: Create snapshot for "head-45bdb83d-vd-root-automatic" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:201 @ 10/01/25 11:05:38.068 STEP: Create snapshot for "head-45bdb83d-vd-root-automatic-with-hotplug" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:201 @ 10/01/25 11:05:45.653 STEP: Create snapshot for "head-45bdb83d-vd-attach-automatic-with-hotplug" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:201 @ 10/01/25 11:05:51.942 < Exit [It] creates snapshots with `requiredConsistency` of attached VDs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:180 @ 10/01/25 11:05:54.139 (18.028s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:05:54.139 < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:05:54.139 (0s) + + + > Enter [It] creates `vdSnapshots` concurrently - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:224 @ 10/01/25 11:05:54.139 STEP: Create five snapshots for "head-45bdb83d-vd-root-automatic-with-hotplug" of "head-45bdb83d-vm-automatic-with-hotplug" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:248 @ 10/01/25 11:06:02.791 STEP: Create five snapshots for "head-45bdb83d-vd-attach-automatic-with-hotplug" of "head-45bdb83d-vm-automatic-with-hotplug" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:248 @ 10/01/25 11:06:07.089 < Exit [It] creates `vdSnapshots` concurrently - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:224 @ 10/01/25 11:06:09.294 (15.15s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:06:09.294 < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:06:09.294 (0s) + + + > Enter [It] checks snapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:285 @ 10/01/25 11:06:09.294 STEP: Snapshots should be `Ready` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:286 @ 10/01/25 11:06:09.294 < Exit [It] checks snapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:285 @ 10/01/25 11:06:28.402 (19.107s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:06:28.402 < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:06:28.402 (0s) + + + > Enter [It] checks `FileSystemFrozen` status of VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:331 @ 10/01/25 11:06:28.402 STEP: Status should not be `Frozen` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:332 @ 10/01/25 11:06:28.402 < Exit [It] checks `FileSystemFrozen` status of VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:331 @ 10/01/25 11:06:31.491 (3.088s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:06:31.491 < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:06:31.491 (0s) + + + > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:357 @ 10/01/25 11:06:31.491 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 11:06:31.491 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 11:06:51.882 (20.39s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:357 @ 10/01/25 11:06:51.882 (20.39s) > Enter [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:06:51.882 < Exit [AfterEach] VirtualDiskSnapshots - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vd_snapshots_test.go:78 @ 10/01/25 11:06:51.882 (0s) + + + > Enter [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/01/25 11:06:51.882 < Exit [BeforeAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:47 @ 10/01/25 11:06:54.163 (2.281s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:67 @ 10/01/25 11:06:54.163 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:67 @ 10/01/25 11:06:58.587 (4.424s) + + + > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/01/25 11:06:58.588 STEP: Virtual machine should be running - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:91 @ 10/01/25 11:06:58.588 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:90 @ 10/01/25 11:08:12.705 (1m14.117s) + + + [FAILED] Timed out after 90.001s. The function passed to Eventually returned the following error: <*errors.errorString | 0x1400071cc20>: cmd: d8 v ssh head-45bdb83d-vm-bar -c 'ping -c 2 -W 2 -w 5 -q 192.168.1.11 2>&1 | grep -o "[0-9]\+%\s*packet loss"' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-vm-vpc --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed stderr: bash: d8: command not found { s: "cmd: d8 v ssh head-45bdb83d-vm-bar -c 'ping -c 2 -W 2 -w 5 -q 192.168.1.11 2>&1 | grep -o \"[0-9]\\+%\\s*packet loss\"' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-vm-vpc --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed\nstderr: bash: d8: command not found\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:213 @ 10/01/25 11:09:46.019 + > Enter [It] checks network availability - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/01/25 11:08:12.706 STEP: Network condition should be true - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:99 @ 10/01/25 11:08:12.706 STEP: VM "head-45bdb83d-vm-bar" should have connectivity to 192.168.1.11 (target: vm-foo) - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:212 @ 10/01/25 11:08:16.017 [FAILED] Timed out after 90.001s. The function passed to Eventually returned the following error: <*errors.errorString | 0x1400071cc20>: cmd: d8 v ssh head-45bdb83d-vm-bar -c 'ping -c 2 -W 2 -w 5 -q 192.168.1.11 2>&1 | grep -o "[0-9]\+%\s*packet loss"' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-vm-vpc --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed stderr: bash: d8: command not found { s: "cmd: d8 v ssh head-45bdb83d-vm-bar -c 'ping -c 2 -W 2 -w 5 -q 192.168.1.11 2>&1 | grep -o \"[0-9]\\+%\\s*packet loss\"' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-vm-vpc --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed\nstderr: bash: d8: command not found\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:213 @ 10/01/25 11:09:46.019 < Exit [It] checks network availability - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:98 @ 10/01/25 11:09:46.019 (1m33.313s) > Enter [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/01/25 11:09:46.019 < Exit [AfterAll] VirtualMachineAdditionalNetworkInterfaces - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:60 @ 10/01/25 11:09:51.808 (5.788s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:111 @ 10/01/25 11:09:51.808 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:125 @ 10/01/25 11:09:51.808 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:141 @ 10/01/25 11:09:51.808 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:156 @ 10/01/25 11:09:51.808 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_vpc_test.go:169 @ 10/01/25 11:09:51.808 + + + > Enter [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/01/25 11:09:51.809 < Exit [BeforeAll] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:42 @ 10/01/25 11:09:53.546 (1.737s) > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:09:53.546 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:09:53.546 (0s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/01/25 11:09:53.546 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:64 @ 10/01/25 11:09:56.48 (2.934s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:09:56.48 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:09:56.48 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:09:56.48 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:09:56.48 (0s) > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/01/25 11:09:56.48 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:75 @ 10/01/25 11:09:56.48 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:74 @ 10/01/25 11:10:02.231 (5.751s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:10:02.231 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:10:02.231 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:10:02.232 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:10:02.232 (0s) > Enter [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/01/25 11:10:02.232 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:86 @ 10/01/25 11:10:02.232 < Exit [It] checks VDs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:85 @ 10/01/25 11:10:14.68 (12.448s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:10:14.68 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:10:14.68 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:10:14.68 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:10:14.68 (0s) > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/01/25 11:10:14.68 STEP: Virtual machine phase should be Running - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:97 @ 10/01/25 11:10:14.68 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:96 @ 10/01/25 11:11:02.683 (48.002s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:02.683 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:02.683 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:02.683 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:02.683 (0s) > Enter [It] marks VMs with label map["specialKey":"specialValue"] - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/01/25 11:11:02.683 < Exit [It] marks VMs with label map["specialKey":"specialValue"] - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:107 @ 10/01/25 11:11:04.88 (2.197s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:04.88 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:04.881 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:04.881 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:04.881 (0s) > Enter [It] checks VMs and pods labels after VMs labeling - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/01/25 11:11:04.881 < Exit [It] checks VMs and pods labels after VMs labeling - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:120 @ 10/01/25 11:11:07.143 (2.263s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:07.143 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:07.143 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:07.144 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:07.144 (0s) > Enter [It] removes label map[specialKey:specialValue] from VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/01/25 11:11:07.144 < Exit [It] removes label map[specialKey:specialValue] from VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:152 @ 10/01/25 11:11:09.821 (2.678s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:09.821 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:09.821 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:09.821 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:09.821 (0s) > Enter [It] checks VMs and pods labels after VMs unlabeling - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/01/25 11:11:09.821 < Exit [It] checks VMs and pods labels after VMs unlabeling - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:165 @ 10/01/25 11:11:11.63 (1.809s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:11.63 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:11.63 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:11.631 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:11.631 (0s) > Enter [It] marks VMs with annotation map["specialKey":"specialValue"] - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/01/25 11:11:11.631 < Exit [It] marks VMs with annotation map["specialKey":"specialValue"] - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:199 @ 10/01/25 11:11:14.32 (2.69s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:14.32 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:14.32 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:14.321 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:14.321 (0s) > Enter [It] checks VMs and pods annotations after VMs annotating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/01/25 11:11:14.321 < Exit [It] checks VMs and pods annotations after VMs annotating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:212 @ 10/01/25 11:11:16.075 (1.755s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:16.075 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:16.075 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:16.076 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:16.076 (0s) > Enter [It] removes annotation map[specialKey:specialValue] from VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/01/25 11:11:16.076 < Exit [It] removes annotation map[specialKey:specialValue] from VMs - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:244 @ 10/01/25 11:11:18.253 (2.177s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:18.253 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:18.253 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:18.253 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:18.253 (0s) > Enter [It] checks VMs and pods annotations after VMs unannotating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/01/25 11:11:18.253 < Exit [It] checks VMs and pods annotations after VMs unannotating - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:257 @ 10/01/25 11:11:20.053 (1.8s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:20.053 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:11:20.053 (0s) + + + > Enter [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:20.054 < Exit [BeforeEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:51 @ 10/01/25 11:11:20.054 (0s) > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/01/25 11:11:20.054 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 11:11:20.054 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 11:12:05.899 (45.845s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:291 @ 10/01/25 11:12:05.899 (45.845s) > Enter [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:12:05.899 < Exit [AfterEach] VirtualMachineLabelAndAnnotation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_label_annotation_test.go:57 @ 10/01/25 11:12:05.899 (0s) + + + > Enter [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/01/25 11:12:05.9 < Exit [BeforeAll] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:50 @ 10/01/25 11:12:08.838 (2.938s) > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/01/25 11:12:08.838 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/01/25 11:12:08.838 (0s) > Enter [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/01/25 11:12:08.838 STEP: Create a vmip automatically and check its binding with a lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:81 @ 10/01/25 11:12:08.838 STEP: Remove label from the lease - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:87 @ 10/01/25 11:12:09.724 STEP: Wait for the label to be restored by the controller - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:96 @ 10/01/25 11:12:09.864 < Exit [It] Creates vmip with type Auto - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:80 @ 10/01/25 11:12:10.137 (1.299s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/01/25 11:12:10.137 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/01/25 11:12:10.137 (0s) + + + > Enter [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/01/25 11:12:10.137 < Exit [BeforeEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:65 @ 10/01/25 11:12:10.137 (0s) > Enter [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/01/25 11:12:10.137 STEP: Create an intermediate vmip automatically to allocate a new ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:109 @ 10/01/25 11:12:10.137 STEP: Delete the intermediate vmip automatically and check that the lease is released - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:115 @ 10/01/25 11:12:10.54 STEP: Reuse the released lease with a static vmip - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:124 @ 10/01/25 11:12:10.805 STEP: Delete the static vmip and lease, then create another static vmip with this ip address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:132 @ 10/01/25 11:12:11.215 < Exit [It] Creates vmip with type Static - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:108 @ 10/01/25 11:12:11.954 (1.817s) > Enter [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/01/25 11:12:11.954 < Exit [AfterEach] IPAM - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/ipam_test.go:157 @ 10/01/25 11:12:11.954 (0s) + + + > Enter [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/01/25 11:12:11.954 < Exit [BeforeAll] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:36 @ 10/01/25 11:12:13.699 (1.744s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/01/25 11:12:13.699 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:52 @ 10/01/25 11:12:19.742 (6.043s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/01/25 11:12:19.742 < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/01/25 11:12:19.742 (0s) + + + > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/01/25 11:12:19.742 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:76 @ 10/01/25 11:12:19.742 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:75 @ 10/01/25 11:13:58.107 (1m38.364s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/01/25 11:13:58.107 < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/01/25 11:13:58.107 (0s) + + + > Enter [It] starts migrations - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:86 @ 10/01/25 11:13:58.107 < Exit [It] starts migrations - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:86 @ 10/01/25 11:14:00.014 (1.907s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/01/25 11:14:00.014 < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/01/25 11:14:00.014 (0s) + + + > Enter [It] checks VMs and VMOPs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:100 @ 10/01/25 11:14:00.014 STEP: VMOPs should be in Completed phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:101 @ 10/01/25 11:14:00.014 STEP: Virtual machines should be migrated - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:107 @ 10/01/25 11:14:23.518 < Exit [It] checks VMs and VMOPs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:100 @ 10/01/25 11:14:25.84 (25.826s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/01/25 11:14:25.84 < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/01/25 11:14:25.84 (0s) + + + [FAILED] Timed out after 90.001s. The function passed to Eventually returned the following error: <*errors.errorString | 0x1400098aa00>: cmd: d8 v ssh head-45bdb83d-vm-migration-bios -c 'curl -o /dev/null -s -w "%{http_code}\n" https://flant.com' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-vm-migration --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed stderr: bash: d8: command not found { s: "cmd: d8 v ssh head-45bdb83d-vm-migration-bios -c 'curl -o /dev/null -s -w \"%{http_code}\\n\" https://flant.com' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-vm-migration --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed\nstderr: bash: d8: command not found\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:126 @ 10/01/25 11:16:29.551 + > Enter [It] checks VMs external connection after migrations - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:116 @ 10/01/25 11:14:25.84 STEP: Cilium agent should be OK's for VM: head-45bdb83d-vm-migration-bios - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:125 @ 10/01/25 11:14:26.776 STEP: Cilium agent should be OK's for VM: head-45bdb83d-vm-migration-uefi - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:125 @ 10/01/25 11:14:43.434 STEP: Response code from "https://flant.com" should be "200" for "head-45bdb83d-vm-migration-bios" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:126 @ 10/01/25 11:14:59.548 [FAILED] Timed out after 90.001s. The function passed to Eventually returned the following error: <*errors.errorString | 0x1400098aa00>: cmd: d8 v ssh head-45bdb83d-vm-migration-bios -c 'curl -o /dev/null -s -w "%{http_code}\n" https://flant.com' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-vm-migration --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed stderr: bash: d8: command not found { s: "cmd: d8 v ssh head-45bdb83d-vm-migration-bios -c 'curl -o /dev/null -s -w \"%{http_code}\\n\" https://flant.com' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-vm-migration --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed\nstderr: bash: d8: command not found\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:126 @ 10/01/25 11:16:29.551 < Exit [It] checks VMs external connection after migrations - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:116 @ 10/01/25 11:16:29.551 (2m3.71s) > Enter [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/01/25 11:16:29.551 < Exit [AfterEach] VirtualMachineMigration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:45 @ 10/01/25 11:16:47.599 (18.048s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_migration_test.go:131 @ 10/01/25 11:16:47.599 + + + > Enter [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:51 @ 10/01/25 11:16:47.599 < Exit [BeforeAll] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:51 @ 10/01/25 11:16:52.255 (4.656s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:61 @ 10/01/25 11:16:52.255 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:61 @ 10/01/25 11:17:44.741 (52.485s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:17:44.741 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:17:44.741 (0s) + + + > Enter [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:82 @ 10/01/25 11:17:44.741 < Exit [It] should fill empty virtualMachineClassName with the default class name - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:82 @ 10/01/25 11:17:45.914 (1.173s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:17:45.914 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:17:45.914 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:97 @ 10/01/25 11:17:45.914 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:98 @ 10/01/25 11:17:45.914 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:97 @ 10/01/25 11:17:46.766 (852ms) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:17:46.766 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:17:46.766 (0s) + + + > Enter [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:108 @ 10/01/25 11:17:46.766 STEP: CVIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:109 @ 10/01/25 11:17:46.766 < Exit [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:108 @ 10/01/25 11:17:47.975 (1.209s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:17:47.975 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:17:47.975 (0s) + + + > Enter [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:119 @ 10/01/25 11:17:47.976 STEP: VMClasses should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:120 @ 10/01/25 11:17:47.976 < Exit [It] checks VMClasses phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:119 @ 10/01/25 11:17:49.204 (1.228s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:17:49.204 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:17:49.204 (0s) + + + > Enter [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 @ 10/01/25 11:17:49.204 < Exit [It] patches custom VMIP with unassigned address - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:130 @ 10/01/25 11:17:54.566 (5.362s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:17:54.566 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:17:54.566 (0s) + + + > Enter [It] checks VMIPs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:137 @ 10/01/25 11:17:54.566 STEP: VMIPs should be in Attached phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:138 @ 10/01/25 11:17:54.566 < Exit [It] checks VMIPs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:137 @ 10/01/25 11:17:55.624 (1.058s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:17:55.624 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:17:55.624 (0s) + + + > Enter [It] checks VDs phases with consumers - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:148 @ 10/01/25 11:17:55.624 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:149 @ 10/01/25 11:17:55.624 < Exit [It] checks VDs phases with consumers - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:148 @ 10/01/25 11:19:02.855 (1m7.231s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:19:02.855 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:19:02.855 (0s) + + + > Enter [It] checks VDs phases with no consumers - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:158 @ 10/01/25 11:19:02.856 STEP: VDs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:159 @ 10/01/25 11:19:02.856 < Exit [It] checks VDs phases with no consumers - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:158 @ 10/01/25 11:19:06.344 (3.488s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:19:06.344 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:19:06.344 (0s) + + + > Enter [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:169 @ 10/01/25 11:19:06.344 STEP: Virtual machine agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:170 @ 10/01/25 11:19:06.344 < Exit [It] checks VMs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:169 @ 10/01/25 11:19:56.819 (50.475s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:19:56.819 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:19:56.819 (0s) + + + > Enter [It] checks VMBDAs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:180 @ 10/01/25 11:19:56.819 STEP: VMBDAs should be in Attached phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:181 @ 10/01/25 11:19:56.819 < Exit [It] checks VMBDAs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:180 @ 10/01/25 11:19:57.947 (1.127s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:19:57.947 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:19:57.947 (0s) + + + [FAILED] Timed out after 90.001s. The function passed to Eventually returned the following error: <*errors.errorString | 0x1400158e280>: cmd: d8 v ssh head-45bdb83d-vm-default -c 'curl -o /dev/null -s -w "%{http_code}\n" https://flant.com' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-complex-test --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed stderr: bash: d8: command not found { s: "cmd: d8 v ssh head-45bdb83d-vm-default -c 'curl -o /dev/null -s -w \"%{http_code}\\n\" https://flant.com' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-complex-test --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed\nstderr: bash: d8: command not found\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:202 @ 10/01/25 11:21:45.832 + > Enter [It] checks VMs external connectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:192 @ 10/01/25 11:19:57.947 STEP: Cilium agent should be OK's for VM: head-45bdb83d-vm-default - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:201 @ 10/01/25 11:19:58.866 STEP: Response code from "https://flant.com" should be "200" for "head-45bdb83d-vm-default" - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:202 @ 10/01/25 11:20:15.882 [FAILED] Timed out after 90.001s. The function passed to Eventually returned the following error: <*errors.errorString | 0x1400158e280>: cmd: d8 v ssh head-45bdb83d-vm-default -c 'curl -o /dev/null -s -w "%{http_code}\n" https://flant.com' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-complex-test --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed stderr: bash: d8: command not found { s: "cmd: d8 v ssh head-45bdb83d-vm-default -c 'curl -o /dev/null -s -w \"%{http_code}\\n\" https://flant.com' --local-ssh=true --local-ssh-opts='-o StrictHostKeyChecking=no' --local-ssh-opts='-o UserKnownHostsFile=/dev/null' --local-ssh-opts='-o LogLevel=ERROR' --local-ssh-opts='-o ConnectTimeout=10s' -n head-45bdb83d-end-to-end-complex-test --username=cloud --identity-file=/tmp/testdata/sshkeys/id_ed\nstderr: bash: d8: command not found\n", } In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:202 @ 10/01/25 11:21:45.832 < Exit [It] checks VMs external connectivity - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:192 @ 10/01/25 11:21:45.832 (1m47.936s) > Enter [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:21:45.832 < Exit [AfterEach] ComplexTest - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:45 @ 10/01/25 11:21:51.565 (5.732s) + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:216 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:240 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:263 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:275 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:293 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:310 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:334 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:347 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:351 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:368 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:381 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:398 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:437 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:472 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:490 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:505 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:521 @ 10/01/25 11:21:51.565 + + + + [SKIPPED] Spec skipped because an earlier spec in an ordered container failed In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/complex_test.go:538 @ 10/01/25 11:21:51.565 + + + > Enter [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/01/25 11:21:51.565 < Exit [BeforeAll] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:59 @ 10/01/25 11:21:53.323 (1.758s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/01/25 11:21:53.323 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:79 @ 10/01/25 11:22:03.031 (9.708s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/01/25 11:22:03.031 < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/01/25 11:22:03.031 (0s) + + + > Enter [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/01/25 11:22:03.031 STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/01/25 11:22:03.032 END STEP: `VirtualImages` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:88 @ 10/01/25 11:22:10.398 (7.367s) STEP: `VirtualMachineClasses` should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:95 @ 10/01/25 11:22:10.398 END STEP: `VirtualMachineClasses` should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:95 @ 10/01/25 11:22:12.602 (2.204s) STEP: `VirtualDisks` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:102 @ 10/01/25 11:22:12.602 END STEP: `VirtualDisks` should be in the "Ready" phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:102 @ 10/01/25 11:23:08.927 (56.325s) STEP: `VirtualMachines` agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:109 @ 10/01/25 11:23:08.927 END STEP: `VirtualMachines` agents should be ready - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:109 @ 10/01/25 11:24:27.267 (1m18.34s) < Exit [It] checks the resources phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:87 @ 10/01/25 11:24:27.267 (2m24.236s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/01/25 11:24:27.267 < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/01/25 11:24:27.267 (0s) + + + > Enter [It] checks the `status.nodeName` field of the `VirtualMachines` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:120 @ 10/01/25 11:24:27.267 STEP: Obtain the `VirtualMachine` objects - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:128 @ 10/01/25 11:24:27.267 END STEP: Obtain the `VirtualMachine` objects - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:128 @ 10/01/25 11:24:38.498 (11.231s) STEP: Set affinity when creating the `VirtualMachines`.`: `vm-a` and `vm-c` should be running on the same node - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:138 @ 10/01/25 11:24:38.498 END STEP: Set affinity when creating the `VirtualMachines`.`: `vm-a` and `vm-c` should be running on the same node - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:138 @ 10/01/25 11:24:38.498 (0s) STEP: Set anti-affinity when creating the `VirtualMachines`: `vm-a` and `vm-b` should be running on the different nodes - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:141 @ 10/01/25 11:24:38.498 END STEP: Set anti-affinity when creating the `VirtualMachines`: `vm-a` and `vm-b` should be running on the different nodes - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:141 @ 10/01/25 11:24:38.498 (0s) STEP: Set toleration when creating the `VirtualMachines`: `vm-d` should be running on a master node - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:144 @ 10/01/25 11:24:38.498 END STEP: Set toleration when creating the `VirtualMachines`: `vm-d` should be running on a master node - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:144 @ 10/01/25 11:24:39.295 (796ms) STEP: Change affinity to anti-affinity when the `VirtualMachines` are runnning: `vm-a` and `vm-c` should be running on the different nodes - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:150 @ 10/01/25 11:24:39.295 END STEP: Change affinity to anti-affinity when the `VirtualMachines` are runnning: `vm-a` and `vm-c` should be running on the different nodes - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:150 @ 10/01/25 11:24:53.941 (14.647s) STEP: Change anti-affinity to affinity when the `VirtualMachines` are runnning: `vm-a` and `vm-c` should be running on the same node - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:207 @ 10/01/25 11:24:53.941 END STEP: Change anti-affinity to affinity when the `VirtualMachines` are runnning: `vm-a` and `vm-c` should be running on the same node - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:207 @ 10/01/25 11:25:22.495 (28.553s) < Exit [It] checks the `status.nodeName` field of the `VirtualMachines` - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:120 @ 10/01/25 11:25:22.495 (55.228s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/01/25 11:25:22.495 < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/01/25 11:25:22.495 (0s) + + + > Enter [It] sets the `spec.nodeSelector` field - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:283 @ 10/01/25 11:25:22.495 STEP: Sets the `spec.nodeSelector` with the `status.nodeSelector` value - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:290 @ 10/01/25 11:25:22.495 END STEP: Sets the `spec.nodeSelector` with the `status.nodeSelector` value - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:290 @ 10/01/25 11:25:24.644 (2.149s) STEP: The `VirtualMachine` should not be migrated - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:300 @ 10/01/25 11:25:24.644 END STEP: The `VirtualMachine` should not be migrated - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:300 @ 10/01/25 11:25:45.562 (20.918s) STEP: Sets the `spec.nodeSelector` with `another node` value - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:315 @ 10/01/25 11:25:45.562 END STEP: Sets the `spec.nodeSelector` with `another node` value - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:315 @ 10/01/25 11:25:54.954 (9.393s) STEP: The `VirtualMachine` should be migrated - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:353 @ 10/01/25 11:25:54.954 END STEP: The `VirtualMachine` should be migrated - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:353 @ 10/01/25 11:26:00.827 (5.872s) < Exit [It] sets the `spec.nodeSelector` field - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:283 @ 10/01/25 11:26:00.827 (38.332s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/01/25 11:26:00.827 < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/01/25 11:26:00.827 (0s) + + + > Enter [It] sets the `spec.affinity.nodeAffinity` field - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:372 @ 10/01/25 11:26:00.827 STEP: Sets the `spec.affinity.nodeAffinity` with the `status.nodeSelector` value - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:379 @ 10/01/25 11:26:00.827 END STEP: Sets the `spec.affinity.nodeAffinity` with the `status.nodeSelector` value - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:379 @ 10/01/25 11:26:02.973 (2.146s) STEP: The `VirtualMachine` should not be migrated - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:392 @ 10/01/25 11:26:02.973 END STEP: The `VirtualMachine` should not be migrated - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:392 @ 10/01/25 11:26:23.561 (20.589s) STEP: Sets the `spec.affinity.nodeAffinity` with `another node` value - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:407 @ 10/01/25 11:26:23.561 END STEP: Sets the `spec.affinity.nodeAffinity` with `another node` value - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:407 @ 10/01/25 11:26:36.808 (13.247s) STEP: The `VirtualMachine` should be migrated - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:448 @ 10/01/25 11:26:36.808 END STEP: The `VirtualMachine` should be migrated - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:448 @ 10/01/25 11:26:42.356 (5.548s) < Exit [It] sets the `spec.affinity.nodeAffinity` field - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:372 @ 10/01/25 11:26:42.356 (41.53s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/01/25 11:26:42.356 < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/01/25 11:26:42.356 (0s) + + + > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:467 @ 10/01/25 11:26:42.357 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 11:26:42.357 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 11:26:59.839 (17.483s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:467 @ 10/01/25 11:26:59.839 (17.483s) > Enter [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/01/25 11:26:59.839 < Exit [AfterEach] VirtualMachineAffinityAndToleration - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/affinity_toleration_test.go:72 @ 10/01/25 11:26:59.839 (0s) + + + > Enter [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:38 @ 10/01/25 11:26:59.84 < Exit [BeforeAll] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:38 @ 10/01/25 11:27:01.578 (1.738s) > Enter [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:77 @ 10/01/25 11:27:01.578 < Exit [It] result should be succeeded - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:77 @ 10/01/25 11:27:21.703 (20.125s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/01/25 11:27:21.703 < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/01/25 11:27:21.703 (0s) + + + > Enter [It] checks VD phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:87 @ 10/01/25 11:27:21.704 STEP: VD should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:88 @ 10/01/25 11:27:21.704 < Exit [It] checks VD phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:87 @ 10/01/25 11:27:38.337 (16.633s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/01/25 11:27:38.337 < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/01/25 11:27:38.337 (0s) + + + > Enter [It] checks VDSnapshot phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:96 @ 10/01/25 11:27:38.338 STEP: VDSnapshot should be in Ready phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:97 @ 10/01/25 11:27:38.338 < Exit [It] checks VDSnapshot phase - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:96 @ 10/01/25 11:27:41.114 (2.776s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/01/25 11:27:41.114 < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/01/25 11:27:41.114 (0s) + + + > Enter [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:107 @ 10/01/25 11:27:41.114 STEP: VIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:108 @ 10/01/25 11:27:41.114 < Exit [It] checks VIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:107 @ 10/01/25 11:28:21.082 (39.967s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/01/25 11:28:21.082 < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/01/25 11:28:21.082 (0s) + + + > Enter [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:116 @ 10/01/25 11:28:21.082 STEP: CVIs should be in Ready phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:117 @ 10/01/25 11:28:21.082 < Exit [It] checks CVIs phases - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:116 @ 10/01/25 11:28:26.093 (5.011s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/01/25 11:28:26.093 < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/01/25 11:28:26.093 (0s) + + + > Enter [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:127 @ 10/01/25 11:28:26.094 STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 11:28:26.094 END STEP: Response on deletion request should be successful - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/util_test.go:604 @ 10/01/25 11:28:41.403 (15.31s) < Exit [It] deletes test case resources - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:127 @ 10/01/25 11:28:41.403 (15.31s) > Enter [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/01/25 11:28:41.403 < Exit [AfterEach] VirtualImageCreation - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/images_creation_test.go:70 @ 10/01/25 11:28:41.403 (0s) + + + + > Enter [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/01/25 11:28:41.404 [SKIPPED] This test case is not working everytime. Should be fixed. In [BeforeAll] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:51 @ 10/01/25 11:28:41.404 < Exit [BeforeAll] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:47 @ 10/01/25 11:28:41.404 (0s) > Enter [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/01/25 11:28:41.404 < Exit [AfterEach] VirtualDiskResizing - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:65 @ 10/01/25 11:28:41.404 (0s) + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:82 @ 10/01/25 11:28:41.404 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:94 @ 10/01/25 11:28:41.404 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:106 @ 10/01/25 11:28:41.404 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:116 @ 10/01/25 11:28:41.404 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:132 @ 10/01/25 11:28:41.405 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:151 @ 10/01/25 11:28:41.405 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:157 @ 10/01/25 11:28:41.405 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:189 @ 10/01/25 11:28:41.405 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:216 @ 10/01/25 11:28:41.405 + + + + [SKIPPED] Spec skipped because Skip() was called in BeforeAll In [It] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/vm_disk_resizing_test.go:252 @ 10/01/25 11:28:41.405 + + + [FAILED] Expected success, but got an error: <*errors.joinError | 0x14000538dc8>: http2: client connection lost http2: client connection lost http2: client connection lost { errs: [ <*errors.joinError | 0x14000710450>{ errs: [ <*errors.joinError | 0x14000710438>{ errs: [ <*errors.errorString | 0x14000448c40>{ s: "http2: client connection lost", }, ], }, <*errors.errorString | 0x14000448c40>{ s: "http2: client connection lost", }, ], }, <*errors.errorString | 0x14000448c40>{ s: "http2: client connection lost", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:184 @ 10/01/25 11:28:42.42 + > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/01/25 11:28:41.405 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/01/25 11:28:41.405 (0s) > Enter [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/01/25 11:28:41.405 [FAILED] Expected success, but got an error: <*errors.joinError | 0x14000538dc8>: http2: client connection lost http2: client connection lost http2: client connection lost { errs: [ <*errors.joinError | 0x14000710450>{ errs: [ <*errors.joinError | 0x14000710438>{ errs: [ <*errors.errorString | 0x14000448c40>{ s: "http2: client connection lost", }, ], }, <*errors.errorString | 0x14000448c40>{ s: "http2: client connection lost", }, ], }, <*errors.errorString | 0x14000448c40>{ s: "http2: client connection lost", }, ], } In [SynchronizedAfterSuite] at: /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:184 @ 10/01/25 11:28:42.42 < Exit [SynchronizedAfterSuite] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:182 @ 10/01/25 11:28:42.42 (1.015s) + + + > Enter [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/01/25 11:28:42.421 < Exit [DeferCleanup (Suite)] TOP-LEVEL - /Users/antont/ansible_deckhouse/virtualization-full/tests/e2e/e2e_test.go:175 @ 10/01/25 11:29:02.357 (19.936s) + + + \ No newline at end of file diff --git a/tests/e2e/artifacts/smoke-sds-20251008-155842/junit.xml b/tests/e2e/artifacts/smoke-sds-20251008-155842/junit.xml new file mode 100644 index 0000000000..715e5a619b --- /dev/null +++ b/tests/e2e/artifacts/smoke-sds-20251008-155842/junit.xml @@ -0,0 +1,658 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + \ No newline at end of file