lauralorenz

lauralorenz

Member Since 8 years ago

Washington, DC

Experience Points
89
follower
Lessons Completed
7
follow
Lessons Completed
39
stars
Best Reply Awards
53
repos

53 contributions in the last year

Pinned
⚡ Airflow plugin to transfer arbitrary files between operators
Activity
May
19
4 days ago
Activity icon
issue

lauralorenz issue kubernetes-sigs/about-api

lauralorenz
lauralorenz

Customize kubectl get

Add a custom value column to the ClusterProperty CRD, so basic kubectl get shows a truncated value

May
2
3 weeks ago
Activity icon
issue

lauralorenz issue comment kubernetes/enhancements

lauralorenz
lauralorenz

ClusterID for ClusterSet identification

Enhancement Description

Please keep this description up to date. This will help the Enhancement Team to track the evolution of the enhancement efficiently.

lauralorenz
lauralorenz
Apr
18
1 month ago
Apr
14
1 month ago
push

lauralorenz push lauralorenz/multicluster

lauralorenz
lauralorenz

parity with new function signatures

Signed-off-by: lauralorenz [email protected]

commit sha: 970653f3ef74e14c16216e8362983ca5019c4406

push time in 1 month ago
pull request

lauralorenz pull request coredns/multicluster

lauralorenz
lauralorenz

parity with new function signatures fixes #7

fixes #7

Activity icon
created branch

lauralorenz in lauralorenz/multicluster create branch 7-update-for-coredns-191

createdAt 1 month ago
Activity icon
issue

lauralorenz issue coredns/multicluster

lauralorenz
lauralorenz

Update for parity with changes in record functions

Plugin does not build with coredns main branch, I tried at coredns commit a929b0b1ecbb0f87b0bd0cd0a40ce38893dce9c0

[email protected]:~/coredns/coredns$ make
CGO_ENABLED=0  go build -v -ldflags="-s -w -X github.com/coredns/coredns/coremain.GitCommit=a929b0b1-dirty" -o coredns
github.com/coredns/multicluster
# github.com/coredns/multicluster
../../go/pkg/mod/github.com/coredns/[email protected]/multicluster.go:131:18: assignment mismatch: 2 variables but plugin.A returns 3 values
../../go/pkg/mod/github.com/coredns/[email protected]/multicluster.go:133:18: assignment mismatch: 2 variables but plugin.AAAA returns 3 values
../../go/pkg/mod/github.com/coredns/[email protected]/multicluster.go:135:18: assignment mismatch: 2 variables but plugin.TXT returns 3 values
../../go/pkg/mod/github.com/coredns/[email protected]/multicluster.go:148:12: assignment mismatch: 2 variables but plugin.A returns 3 values
make: *** [Makefile:16: coredns] Error 2

Appears to be due to function signature changes for records in https://github.com/coredns/coredns/pull/4715 meaning this plugin is currently incompatible as of coredns 1.9.1.

Activity icon
fork

lauralorenz forked coredns/multicluster

⚡ CoreDNS plugin implementing K8s multi-cluster services DNS spec.
lauralorenz Updated
fork time in 1 month ago
Apr
11
1 month ago
Activity icon
issue

lauralorenz issue comment kubernetes/kubernetes

lauralorenz
lauralorenz

[Flaky tests] Several builds intermittently failing outside of test results: gce-cos-master-reboot, gce-cos-master-default, gce-cos-master-alpha-features, gce-cos-master-scalability-100

Which jobs are flaking?

master-blocking/gce-cos-master-reboot master-blocking/gce-cos-master-default master-blocking/gce-cos-master-alpha-features master-blocking/gce-cos-master-scalability-100

Which tests are flaking?

All these tests seem to have similar starting test named "Overall" and "Pod" that are failing, ex ci-kubernetes-e2e-gci-gce-alpha-features.Pod and ci-kubernetes-e2e-gci-gce-alpha-features.Overall

Since when has it been flaking?

4/7, seems to reoccur daily or several times a day even with no change in test commit.

Testgrid link

See above links for each job

Reason for failure (if possible)

gce-cos-master-scalability-100's message is different from the others:

Init container clonerefs not ready: (state: waiting, reason: "PodInitializing", message: "") Init container initupload not ready: (state: waiting, reason: "PodInitializing", message: "") Init container place-entrypoint not ready: (state: waiting, reason: "PodInitializing", message: "")

All the others seem to be more similar (master-blocking/gce-cos-master-reboot, master-blocking/gce-cos-master-default, master-blocking/gce-cos-master-alpha-features) and emit either Job execution failed: Pod pending timeout. or Job execution failed: Pod got deleted unexpectedly; in the pod events the latest events I see are:

Warning | FailedCreatePodSandBox | 1h18m | kubelet | Failed to create pod sandbox: rpc error: code = Unknown desc = failed to reserve sandbox name "e62f5752-b9ae-11ec-afd0-aefa2f608b87_test-pods_da9a67b1-53d5-40ee-a40c-9bfec3582552_0": name "e62f5752-b9ae-11ec-afd0-aefa2f608b87_test-pods_da9a67b1-53d5-40ee-a40c-9bfec3582552_0" is reserved for "3a0532714cdaad3a07206f89fb21262b6e1d0bb8f252372aed0a9a52d70d90e6"

or

Warning | NodeNotReady | 1h52m | node-controller | Node is not ready

Anything else we need to know?

The issue with master-blocking/gce-cos-master-scalability-100 was previously observed by testing-ops on 4/8 (see slack thread) and was determined to resolve on its own then, but has reoccured after that.

Relevant SIG(s)

/sig testing

Activity icon
issue

lauralorenz issue kubernetes/kubernetes

lauralorenz
lauralorenz

[Flaky tests] Several builds intermittently failing outside of test results: gce-cos-master-reboot, gce-cos-master-default, gce-cos-master-alpha-features, gce-cos-master-scalability-100

Which jobs are flaking?

master-blocking/gce-cos-master-reboot master-blocking/gce-cos-master-default master-blocking/gce-cos-master-alpha-features master-blocking/gce-cos-master-scalability-100

Which tests are flaking?

All these tests seem to have similar starting test named "Overall" and "Pod" that are failing, ex ci-kubernetes-e2e-gci-gce-alpha-features.Pod and ci-kubernetes-e2e-gci-gce-alpha-features.Overall

Since when has it been flaking?

4/7, seems to reoccur daily or several times a day even with no change in test commit.

Testgrid link

See above links for each job

Reason for failure (if possible)

gce-cos-master-scalability-100's message is different from the others:

Init container clonerefs not ready: (state: waiting, reason: "PodInitializing", message: "") Init container initupload not ready: (state: waiting, reason: "PodInitializing", message: "") Init container place-entrypoint not ready: (state: waiting, reason: "PodInitializing", message: "")

All the others seem to be more similar (master-blocking/gce-cos-master-reboot, master-blocking/gce-cos-master-default, master-blocking/gce-cos-master-alpha-features) and emit either Job execution failed: Pod pending timeout. or Job execution failed: Pod got deleted unexpectedly; in the pod events the latest events I see are:

Warning | FailedCreatePodSandBox | 1h18m | kubelet | Failed to create pod sandbox: rpc error: code = Unknown desc = failed to reserve sandbox name "e62f5752-b9ae-11ec-afd0-aefa2f608b87_test-pods_da9a67b1-53d5-40ee-a40c-9bfec3582552_0": name "e62f5752-b9ae-11ec-afd0-aefa2f608b87_test-pods_da9a67b1-53d5-40ee-a40c-9bfec3582552_0" is reserved for "3a0532714cdaad3a07206f89fb21262b6e1d0bb8f252372aed0a9a52d70d90e6"

or

Warning | NodeNotReady | 1h52m | node-controller | Node is not ready

Anything else we need to know?

The issue with master-blocking/gce-cos-master-scalability-100 was previously observed by testing-ops on 4/8 (see slack thread) and was determined to resolve on its own then, but has reoccured after that.

Relevant SIG(s)

/sig testing

Apr
4
1 month ago
Activity icon
issue

lauralorenz issue comment kubernetes-sigs/about-api

lauralorenz
lauralorenz

First commit for clusterproperty controller with the CRD

Hello,

Please review this first commit for the clusterproperty controller.

Included in the PR:

  1. Make files to build and deploy the controller v1alpha version.
  2. Create the CRD for clusterproperty scoped to cluster.
  3. Create api group about.k8s.io.
  4. kustomize patched for clusterproperty yaml.
  5. Add the CRD to k8s.io apigroup as explained in this pull request. : https://github.com/kubernetes/enhancements/pull/1100

Thank you,

Apr
1
1 month ago
Activity icon
issue

lauralorenz issue comment kubernetes/kubernetes

lauralorenz
lauralorenz

[Failing test] gce-cos-master-serial

Which jobs are failing?

sig-release-master-informing#gce-cos-master-serial

Which tests are failing?

Kubernetes e2e suite.[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. kubetest.Test ci-kubernetes-e2e-gci-gce-serial.Pod ci-kubernetes-e2e-gci-gce-serial.Overall

Since when has it been failing?

3/31 02:37 PDT

Testgrid link

https://testgrid.k8s.io/sig-release-master-informing#gce-cos-master-serial

Reason for failure (if possible)

Job execution failed: Pod pending timeout.
{ Failure test/e2e/storage/testsuites/disruptive.go:155
Mar 31 11:13:26.559: Encountered SSH error.
Unexpected error:
    <*errors.errorString | 0xc001202e50>: {
        s: "error getting SSH client to [email protected]:22: 'ssh: handshake failed: read tcp 10.34.99.13:46652->34.82.121.164:22: read: connection reset by peer'",
    }
    error getting SSH client to [email protected]:22: 'ssh: handshake failed: read tcp 10.34.99.13:46652->34.82.121.164:22: read: connection reset by peer'
occurred
test/e2e/storage/utils/utils.go:283}

Anything else we need to know?

No response

Relevant SIG(s)

/sig cloud-provider

lauralorenz
lauralorenz

This appears to have recovered for the past 4 runs and exhibited closer to flaky behavior in the first place; will reopen if it goes back into failing state.

/close

Mar
31
1 month ago
Activity icon
issue

lauralorenz issue kubernetes/kubernetes

lauralorenz
lauralorenz

[Failing test] gce-cos-master-serial

Which jobs are failing?

sig-release-master-informing#gce-cos-master-serial

Which tests are failing?

Kubernetes e2e suite.[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] disruptive[Disruptive][LinuxOnly] Should test that pv used in a pod that is deleted while the kubelet is down cleans up when the kubelet returns. kubetest.Test ci-kubernetes-e2e-gci-gce-serial.Pod ci-kubernetes-e2e-gci-gce-serial.Overall

Since when has it been failing?

3/31 02:37 PDT

Testgrid link

https://testgrid.k8s.io/sig-release-master-informing#gce-cos-master-serial

Reason for failure (if possible)

Job execution failed: Pod pending timeout.
{ Failure test/e2e/storage/testsuites/disruptive.go:155
Mar 31 11:13:26.559: Encountered SSH error.
Unexpected error:
    <*errors.errorString | 0xc001202e50>: {
        s: "error getting SSH client to [email protected]:22: 'ssh: handshake failed: read tcp 10.34.99.13:46652->34.82.121.164:22: read: connection reset by peer'",
    }
    error getting SSH client to [email protected]:22: 'ssh: handshake failed: read tcp 10.34.99.13:46652->34.82.121.164:22: read: connection reset by peer'
occurred
test/e2e/storage/utils/utils.go:283}

Anything else we need to know?

No response

Relevant SIG(s)

/sig cloud-provider

Activity icon
issue

lauralorenz issue kubernetes/kubernetes

lauralorenz
lauralorenz

[Failing test] gce-master-scale-performance

Which jobs are failing?

sig-release-master-informing#gce-master-scale-performance

Which tests are failing?

ClusterLoaderV2.huge-service overall (testing/huge-service/config.yaml)[Changes](https://github.com/kubernetes/kubernetes/compare/6c96ac04f...2e55595d3?)ClusterLoaderV2.huge-service: [step: 09] gathering measurements[Changes](https://github.com/kubernetes/kubernetes/compare/6c96ac04f...2e55595d3?)ClusterLoaderV2.huge-service: [step: 09] gathering measurements / Measurement[01] - APIResponsivenessPrometheus - APIResponsivenessPrometheusSimple[Changes](https://github.com/kubernetes/kubernetes/compare/6c96ac04f...2e55595d3?)ClusterLoaderV2.load overall (testing/load/config.yaml)[Changes](https://github.com/kubernetes/kubernetes/compare/6c96ac04f...2e55595d3?)ClusterLoaderV2.load: [step: 19] Collecting pod startup latency[Changes](https://github.com/kubernetes/kubernetes/compare/6c96ac04f...2e55595d3?)ClusterLoaderV2.load: [step: 19] Collecting pod startup latency / Measurement[00] - PodStartupLatency - PodStartupLatency[Changes](https://github.com/kubernetes/kubernetes/compare/6c96ac04f...2e55595d3?)ClusterLoaderV2.load: [step: 22] Waiting for 'scale and update objects' to be completed[Changes](https://github.com/kubernetes/kubernetes/compare/6c96ac04f...2e55595d3?)ClusterLoaderV2.load: [step: 22] Waiting for 'scale and update objects' to be completed / Measurement[00] - WaitForControlledPodsRunning -[Changes](https://github.com/kubernetes/kubernetes/compare/6c96ac04f...2e55595d3?)ClusterLoaderV2.load: [step: 25] Waiting for 'delete objects' to be completed[Changes](https://github.com/kubernetes/kubernetes/compare/6c96ac04f...2e55595d3?)ClusterLoaderV2.load: [step: 25] Waiting for 'delete objects' to be completed / Measurement[00] - WaitForControlledPodsRunning -[Changes](https://github.com/kubernetes/kubernetes/compare/6c96ac04f...2e55595d3?)ClusterLoaderV2.load: [step: 29] gathering measurements[Changes](https://github.com/kubernetes/kubernetes/compare/6c96ac04f...2e55595d3?)ClusterLoaderV2.load: [step: 29] gathering measurements / Measurement[01] - APIResponsivenessPrometheus - APIResponsivenessPrometheusSimple[Changes](https://github.com/kubernetes/kubernetes/compare/6c96ac04f...2e55595d3?)ci-kubernetes-e2e-gce-scale-performance.Overall[Changes](https://github.com/kubernetes/kubernetes/compare/6c96ac04f...2e55595d3?)kubetest.ClusterLoaderV2[Changes](https://github.com/kubernetes/kubernetes/compare/6c96ac04f...2e55595d3?)kubetest.TearDown[Changes](https://github.com/kubernetes/kubernetes/compare/6c96ac04f...2e55595d3?)kubetest.Timeout[Changes](https://github.com/kubernetes/kubernetes/compare/6c96ac04f...2e55595d3?)

Since when has it been failing?

3/30, on 2e55595d3

Testgrid link

https://testgrid.k8s.io/sig-release-master-informing#gce-master-scale-performance

Reason for failure (if possible)



ClusterLoaderV2: huge-service overall (testing/huge-service/config.yaml) expand_less | 2m57s
-- | --
{ Failure :0 [measurement call APIResponsivenessPrometheus - APIResponsivenessPrometheusSimple error: top latency metric: there should be no high-latency requests, but: [got: &{Resource:pods Subresource: Verb:DELETE Scope:resource Latency:perc50: 98.076923ms, perc90: 588.953488ms, perc99: 1.227272727s Count:2087 SlowCount:42}; expected perc99 <= 1s]] :0}

{ Failure :0
[measurement call PodStartupLatency - PodStartupLatency error: pod startup: too high latency 50th percentile: got 54.8916778s expected: 5s
measurement call WaitForControlledPodsRunning - WaitForRunningDaemonSets error: 1 objects timed out: DaemonSets: test-7nniqa-1/daemonset-0
measurement call WaitForControlledPodsRunning - WaitForRunningStatefulSets error: 26 objects timed out:

[... long list of pod names redacted ...]

 expected perc99 <= 1s got: &{Resource:pods Subresource: Verb:GET Scope:resource Latency:perc50: 424.826697ms, perc90: 1.334616169s, perc99: 5.374065701s Count:26420193 SlowCount:4209570}; expected perc99 <= 1s]]

Anything else we need to know?

No response

Relevant SIG(s)

/sig scalability

Mar
30
1 month ago
open pull request

lauralorenz wants to merge kubernetes-sigs/about-api

lauralorenz
lauralorenz

First commit for clusterproperty controller with the CRD

Hello,

Please review this first commit for the clusterproperty controller.

Included in the PR:

  1. Make files to build and deploy the controller v1alpha version.
  2. Create the CRD for clusterproperty scoped to cluster.
  3. Create api group about.k8s.io.
  4. kustomize patched for clusterproperty yaml.
  5. Add the CRD to k8s.io apigroup as explained in this pull request. : https://github.com/kubernetes/enhancements/pull/1100

Thank you,

lauralorenz
lauralorenz

tiny nit:

# using a kube-system ns uuid as the id value:
pull request

lauralorenz merge to kubernetes-sigs/about-api

lauralorenz
lauralorenz

First commit for clusterproperty controller with the CRD

Hello,

Please review this first commit for the clusterproperty controller.

Included in the PR:

  1. Make files to build and deploy the controller v1alpha version.
  2. Create the CRD for clusterproperty scoped to cluster.
  3. Create api group about.k8s.io.
  4. kustomize patched for clusterproperty yaml.
  5. Add the CRD to k8s.io apigroup as explained in this pull request. : https://github.com/kubernetes/enhancements/pull/1100

Thank you,

Activity icon
issue

lauralorenz issue comment kubernetes/kubernetes

lauralorenz
lauralorenz

[Failing test] sig-release-master-blocking#ci-kubernetes-unit

Which jobs are failing?

sig-release-master-blocking#ci-kubernetes-unit

Which tests are failing?

k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.TestPreShutdownHooks

Since when has it been failing?

17:42 PDT 3/29

Testgrid link

https://testgrid.k8s.io/sig-release-master-blocking#ci-kubernetes-unit

Reason for failure (if possible)

{Failed  === RUN   TestPreShutdownHooks
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-1
I0330 02:05:41.123870   78579 tlsconfig.go:240] "Starting DynamicServingCertificateController"
I0330 02:05:41.124641   78579 secure_serving.go:210] Serving securely on [::]:40845
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc00052cf00)}}
    genericapiserver_graceful_termination_test.go:535: Still waiting for the server to start - err: <nil>
    genericapiserver_graceful_termination_test.go:653: [server] received a request, proto: HTTP/2.0, url: /echo?message=attempt-1
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-2
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc0000f1f80)}}
    genericapiserver_graceful_termination_test.go:653: [server] received a request, proto: HTTP/2.0, url: /echo?message=attempt-2
    genericapiserver_graceful_termination_test.go:543: The API server has started
I0330 02:05:41.414509   78579 object_count_tracker.go:84] "StorageObjectCountTracker pruner is exiting"
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-0
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc000436480)}}
    genericapiserver_graceful_termination_test.go:478: PreSHutdown hook can not access the API server - Get "https://127.0.0.1:40845/echo?message=attempt-0": context deadline exceeded
--- FAIL: TestPreShutdownHooks (0.51s)
}

Anything else we need to know?

Failed with the same error for last 4 runs. Beta release cut is scheduled for tomorrow and this is currently a blocker.

Relevant SIG(s)

/sig testing

lauralorenz
lauralorenz

Per [#09QZ4DQB/p1648615574595719?thread_ts=1648607376.354129&cid=C09QZ4DQB) this should be resolved and not be a blocker, testgrid is still red but I think the expectation is that it should clear up. Leaving this open for now but currently expecting this not to block release cut tomorrow.

Activity icon
issue

lauralorenz issue comment kubernetes/kubernetes

lauralorenz
lauralorenz

[Failing test] sig-release-master-blocking#ci-kubernetes-unit

Which jobs are failing?

sig-release-master-blocking#ci-kubernetes-unit

Which tests are failing?

k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.TestPreShutdownHooks

Since when has it been failing?

17:42 PDT 3/29

Testgrid link

https://testgrid.k8s.io/sig-release-master-blocking#ci-kubernetes-unit

Reason for failure (if possible)

{Failed  === RUN   TestPreShutdownHooks
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-1
I0330 02:05:41.123870   78579 tlsconfig.go:240] "Starting DynamicServingCertificateController"
I0330 02:05:41.124641   78579 secure_serving.go:210] Serving securely on [::]:40845
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc00052cf00)}}
    genericapiserver_graceful_termination_test.go:535: Still waiting for the server to start - err: <nil>
    genericapiserver_graceful_termination_test.go:653: [server] received a request, proto: HTTP/2.0, url: /echo?message=attempt-1
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-2
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc0000f1f80)}}
    genericapiserver_graceful_termination_test.go:653: [server] received a request, proto: HTTP/2.0, url: /echo?message=attempt-2
    genericapiserver_graceful_termination_test.go:543: The API server has started
I0330 02:05:41.414509   78579 object_count_tracker.go:84] "StorageObjectCountTracker pruner is exiting"
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-0
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc000436480)}}
    genericapiserver_graceful_termination_test.go:478: PreSHutdown hook can not access the API server - Get "https://127.0.0.1:40845/echo?message=attempt-0": context deadline exceeded
--- FAIL: TestPreShutdownHooks (0.51s)
}

Anything else we need to know?

Failed with the same error for last 4 runs. Beta release cut is scheduled for tomorrow and this is currently a blocker.

Relevant SIG(s)

/sig testing

lauralorenz
lauralorenz

More x-ref: Same post in the [#0EG7JC6T/p1648606081058449).

Activity icon
issue

lauralorenz issue comment kubernetes/kubernetes

lauralorenz
lauralorenz

[Failing test] sig-release-master-blocking#ci-kubernetes-unit

Which jobs are failing?

sig-release-master-blocking#ci-kubernetes-unit

Which tests are failing?

k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.TestPreShutdownHooks

Since when has it been failing?

17:42 PDT 3/29

Testgrid link

https://testgrid.k8s.io/sig-release-master-blocking#ci-kubernetes-unit

Reason for failure (if possible)

{Failed  === RUN   TestPreShutdownHooks
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-1
I0330 02:05:41.123870   78579 tlsconfig.go:240] "Starting DynamicServingCertificateController"
I0330 02:05:41.124641   78579 secure_serving.go:210] Serving securely on [::]:40845
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc00052cf00)}}
    genericapiserver_graceful_termination_test.go:535: Still waiting for the server to start - err: <nil>
    genericapiserver_graceful_termination_test.go:653: [server] received a request, proto: HTTP/2.0, url: /echo?message=attempt-1
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-2
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc0000f1f80)}}
    genericapiserver_graceful_termination_test.go:653: [server] received a request, proto: HTTP/2.0, url: /echo?message=attempt-2
    genericapiserver_graceful_termination_test.go:543: The API server has started
I0330 02:05:41.414509   78579 object_count_tracker.go:84] "StorageObjectCountTracker pruner is exiting"
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-0
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc000436480)}}
    genericapiserver_graceful_termination_test.go:478: PreSHutdown hook can not access the API server - Get "https://127.0.0.1:40845/echo?message=attempt-0": context deadline exceeded
--- FAIL: TestPreShutdownHooks (0.51s)
}

Anything else we need to know?

Failed with the same error for last 4 runs. Beta release cut is scheduled for tomorrow and this is currently a blocker.

Relevant SIG(s)

/sig testing

lauralorenz
lauralorenz

More x-ref: SIG api machinery is tracking it themselves in #109133 There is a PR to increase the timeout in #109134

Activity icon
issue

lauralorenz issue comment kubernetes/kubernetes

lauralorenz
lauralorenz

[Failing test] sig-release-master-blocking#ci-kubernetes-unit

Which jobs are failing?

sig-release-master-blocking#ci-kubernetes-unit

Which tests are failing?

k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.TestPreShutdownHooks

Since when has it been failing?

17:42 PDT 3/29

Testgrid link

https://testgrid.k8s.io/sig-release-master-blocking#ci-kubernetes-unit

Reason for failure (if possible)

{Failed  === RUN   TestPreShutdownHooks
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-1
I0330 02:05:41.123870   78579 tlsconfig.go:240] "Starting DynamicServingCertificateController"
I0330 02:05:41.124641   78579 secure_serving.go:210] Serving securely on [::]:40845
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc00052cf00)}}
    genericapiserver_graceful_termination_test.go:535: Still waiting for the server to start - err: <nil>
    genericapiserver_graceful_termination_test.go:653: [server] received a request, proto: HTTP/2.0, url: /echo?message=attempt-1
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-2
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc0000f1f80)}}
    genericapiserver_graceful_termination_test.go:653: [server] received a request, proto: HTTP/2.0, url: /echo?message=attempt-2
    genericapiserver_graceful_termination_test.go:543: The API server has started
I0330 02:05:41.414509   78579 object_count_tracker.go:84] "StorageObjectCountTracker pruner is exiting"
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-0
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc000436480)}}
    genericapiserver_graceful_termination_test.go:478: PreSHutdown hook can not access the API server - Get "https://127.0.0.1:40845/echo?message=attempt-0": context deadline exceeded
--- FAIL: TestPreShutdownHooks (0.51s)
}

Anything else we need to know?

Failed with the same error for last 4 runs. Beta release cut is scheduled for tomorrow and this is currently a blocker.

Relevant SIG(s)

/sig testing

lauralorenz
lauralorenz
Activity icon
issue

lauralorenz issue comment kubernetes/kubernetes

lauralorenz
lauralorenz

[Failing test] sig-release-master-blocking#ci-kubernetes-unit

Which jobs are failing?

sig-release-master-blocking#ci-kubernetes-unit

Which tests are failing?

k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.TestPreShutdownHooks

Since when has it been failing?

17:42 PDT 3/29

Testgrid link

https://testgrid.k8s.io/sig-release-master-blocking#ci-kubernetes-unit

Reason for failure (if possible)

{Failed  === RUN   TestPreShutdownHooks
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-1
I0330 02:05:41.123870   78579 tlsconfig.go:240] "Starting DynamicServingCertificateController"
I0330 02:05:41.124641   78579 secure_serving.go:210] Serving securely on [::]:40845
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc00052cf00)}}
    genericapiserver_graceful_termination_test.go:535: Still waiting for the server to start - err: <nil>
    genericapiserver_graceful_termination_test.go:653: [server] received a request, proto: HTTP/2.0, url: /echo?message=attempt-1
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-2
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc0000f1f80)}}
    genericapiserver_graceful_termination_test.go:653: [server] received a request, proto: HTTP/2.0, url: /echo?message=attempt-2
    genericapiserver_graceful_termination_test.go:543: The API server has started
I0330 02:05:41.414509   78579 object_count_tracker.go:84] "StorageObjectCountTracker pruner is exiting"
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-0
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc000436480)}}
    genericapiserver_graceful_termination_test.go:478: PreSHutdown hook can not access the API server - Get "https://127.0.0.1:40845/echo?message=attempt-0": context deadline exceeded
--- FAIL: TestPreShutdownHooks (0.51s)
}

Anything else we need to know?

Failed with the same error for last 4 runs. Beta release cut is scheduled for tomorrow and this is currently a blocker.

Relevant SIG(s)

/sig testing

lauralorenz
lauralorenz

Looks like this is also being discussed [#09QZ4DQB/p1648607376354129) and x-ref this test was merged 4 hours ago https://github.com/kubernetes/kubernetes/pull/108033#issuecomment-1082531826

Activity icon
issue

lauralorenz issue kubernetes/kubernetes

lauralorenz
lauralorenz

[Failing test] sig-release-master-blocking#ci-kubernetes-unit

Which jobs are failing?

sig-release-master-blocking#ci-kubernetes-unit

Which tests are failing?

k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.TestPreShutdownHooks

Since when has it been failing?

17:42 PDT 3/29

Testgrid link

https://testgrid.k8s.io/sig-release-master-blocking#ci-kubernetes-unit

Reason for failure (if possible)

{Failed  === RUN   TestPreShutdownHooks
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-1
I0330 02:05:41.123870   78579 tlsconfig.go:240] "Starting DynamicServingCertificateController"
I0330 02:05:41.124641   78579 secure_serving.go:210] Serving securely on [::]:40845
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc00052cf00)}}
    genericapiserver_graceful_termination_test.go:535: Still waiting for the server to start - err: <nil>
    genericapiserver_graceful_termination_test.go:653: [server] received a request, proto: HTTP/2.0, url: /echo?message=attempt-1
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-2
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc0000f1f80)}}
    genericapiserver_graceful_termination_test.go:653: [server] received a request, proto: HTTP/2.0, url: /echo?message=attempt-2
    genericapiserver_graceful_termination_test.go:543: The API server has started
I0330 02:05:41.414509   78579 object_count_tracker.go:84] "StorageObjectCountTracker pruner is exiting"
    genericapiserver_graceful_termination_test.go:560: Sending request - timeout: 100ms, url: https://127.0.0.1:40845/echo?message=attempt-0
    genericapiserver_graceful_termination_test.go:671: [server] seen new connection: &net.TCPConn{conn:net.conn{fd:(*net.netFD)(0xc000436480)}}
    genericapiserver_graceful_termination_test.go:478: PreSHutdown hook can not access the API server - Get "https://127.0.0.1:40845/echo?message=attempt-0": context deadline exceeded
--- FAIL: TestPreShutdownHooks (0.51s)
}

Anything else we need to know?

Failed with the same error for last 4 runs. Beta release cut is scheduled for tomorrow and this is currently a blocker.

Relevant SIG(s)

/sig testing

Mar
28
1 month ago
Activity icon
issue

lauralorenz issue comment kubernetes/kubernetes

lauralorenz
lauralorenz

[Failing test] capz-windows-containerd-master

Which jobs are failing?

capz-windows-containerd-master

Which tests are failing?

Kubernetes e2e suite.BeforeSuite

Since when has it been failing?

2022-03-20 21:01 IST

Testgrid link

https://testgrid.k8s.io/sig-release-master-informing#capz-windows-containerd-master

Reason for failure (if possible)

No response

Anything else we need to know?

No response

Relevant SIG(s)

not sure which sig is responsible here

/sig windows /kind failing-test

cc @kubernetes/ci-signal

lauralorenz
lauralorenz

This is reliably passing now for over 10 runs so closing this issue, will open a new one if behavior changes.

/close

Mar
27
1 month ago
Activity icon
issue

lauralorenz issue comment kubernetes/kubernetes

lauralorenz
lauralorenz

[Flaky test] gci-gce-ingress, diffResource

Which jobs are flaking?

  • gci-gce-ingress

Which tests are flaking?

  • ci-kubernetes-e2e-gci-gce-ingress.Overall
  • kubetest.diffResources

Since when has it been flaking?

History goes back to 2021-12-30

Testgrid link

https://testgrid.k8s.io/sig-release-master-blocking#gci-gce-ingress

Reason for failure (if possible)



1/7062 Tests Failed. | expand_less
-- | --
e2e.go: diffResources expand_less0sError: 2 leaked resources +NAME                              REGION    SESSION_AFFINITY  BACKUP  HEALTH_CHECKS +a1ab9e2204ff44d809dc9bceaaf3c8a7  us-west1  NONE                      a1ab9e2204ff44d809dc9bceaaf3c8a7 | e2e.go: diffResources expand_less | 0s | Error: 2 leaked resources +NAME                              REGION    SESSION_AFFINITY  BACKUP  HEALTH_CHECKS +a1ab9e2204ff44d809dc9bceaaf3c8a7  us-west1  NONE                      a1ab9e2204ff44d809dc9bceaaf3c8a7
e2e.go: diffResources expand_less | 0s
Error: 2 leaked resources +NAME                              REGION    SESSION_AFFINITY  BACKUP  HEALTH_CHECKS +a1ab9e2204ff44d809dc9bceaaf3c8a7  us-west1  NONE                      a1ab9e2204ff44d809dc9bceaaf3c8a7


Anything else we need to know?

No response

Relevant SIG(s)

/sig testing

lauralorenz
lauralorenz

Hi again from CI Signal shadow, x-posting from Slack: boosting this again as this Wednesday 3/30 we plan to cut the 1.24.0-beta.0 release and I would like input as to whether this flake should be considered release blocking since the next candidate is our first (and only) beta candidate.

Mar
26
1 month ago
Activity icon
issue

lauralorenz issue comment kubernetes/kubernetes

lauralorenz
lauralorenz

[Failing test] Conformance-GCE-master-kubetest2, go get not supported

Which jobs are failing?

  • Conformance - GCE - master - kubetest2

Which tests are failing?

  • ci-kubernetes-gce-conformance-latest-kubetest2.Overall

Since when has it been failing?

History goes back to 2022-03-23

Testgrid link

https://testgrid.k8s.io/sig-release-master-blocking#Conformance%20-%20GCE%20-%20master%20-%20kubetest2

Reason for failure (if possible)

+ go get sigs.k8s.io/[email protected]
go: go.mod file not found in current directory or any parent directory.
	'go get' is no longer supported outside a module.
	To build and install a command, use 'go install' with a version,
	like 'go install example.com/[email protected]'
	For more information, see https://golang.org/doc/go-get-install-deprecation
	or run 'go help get' or 'go help install'.
+ EXIT_VALUE=1

Anything else we need to know?

look like a simple replace of go get with go install is needed

Relevant SIG(s)

not sure which sig is responsible here

/sig cloud-provider /kind failing-test

lauralorenz
lauralorenz

Hi, CI Signal rep for the 3/30 release cut here. This test failure is a risk for the 3/30 release cut. Based on the x-ref it looks like it may be related to systemic issues with Go versions (cc @MadhavJivrajani, @liggitt, @BenTheElder, @dims) but I would like someone from the owning SIG, SIG-cloud-provider (cc @andrewsykim, @cheftako, @nckturner) to corroborate.

Mar
22
2 months ago
Activity icon
issue

lauralorenz issue comment kubernetes/enhancements

lauralorenz
lauralorenz

KEP-2149: bump implementable

  • One-line PR description: Bump implementable now that API review and PRR review are both complete
  • Other comments: N/A
pull request

lauralorenz pull request kubernetes/enhancements

lauralorenz
lauralorenz

KEP-2149: bump implementable

  • One-line PR description: Bump implementable now that API review and PRR review are both complete
  • Other comments: N/A
Activity icon
created branch

lauralorenz in lauralorenz/enhancements create branch clusterIDbumpimplementable

createdAt 2 months ago
push

lauralorenz push lauralorenz/enhancements

lauralorenz
lauralorenz

Rename seccomp-operator to security-profiles-operator

Signed-off-by: Sascha Grunert [email protected]

lauralorenz
lauralorenz

mark Disable CAdvisor Json Metrics as implemented

lauralorenz
lauralorenz

add links for cadvisor json metrics disable prs

lauralorenz
lauralorenz
lauralorenz
lauralorenz

Consolidate workload controllers status

lauralorenz
lauralorenz
lauralorenz
lauralorenz

Consolidate workload controllers status: add new conditions + analysis

lauralorenz
lauralorenz

Move KRM function metadata schema to the Catalog KEP

From KEP-2985 to KEP-2906

lauralorenz
lauralorenz
lauralorenz
lauralorenz

Remove support for watch requests from future goals

lauralorenz
lauralorenz

Fix default values for MaxUnavailable and MaxSurge

The implementation defines the default value for MaxSurge being 0, as to preserve the current behavior of rolling updates in DaemonSets. Also, for MaxUnavailable, the rounding is up.

This change is just a copy-paste of the docblock defined in the types.go file.

See:

lauralorenz
lauralorenz

Clarify escaping, root field access and support for int-or-string, embedded and unknown fields

lauralorenz
lauralorenz

Update keps/sig-api-machinery/2876-crd-validation-expression-language/README.md

Co-authored-by: Jordan Liggitt [email protected]

lauralorenz
lauralorenz

Update keps/sig-api-machinery/2876-crd-validation-expression-language/README.md

Co-authored-by: Jordan Liggitt [email protected]

lauralorenz
lauralorenz

Address 'byte' representation and escaping of weird property names

lauralorenz
lauralorenz

Describe how to handle characters not allowed in CEL identifiers in escaping rules

lauralorenz
lauralorenz

Update keps/sig-api-machinery/2876-crd-validation-expression-language/README.md

Co-authored-by: Jordan Liggitt [email protected]

lauralorenz
lauralorenz
lauralorenz
lauralorenz

remove access to object fields as root bound variables

commit sha: a86a0e99ebd9f64d8e3582ad0f9046d6650ae88c

push time in 2 months ago
Mar
16
2 months ago
Activity icon
issue

lauralorenz issue comment kubernetes/kubernetes

lauralorenz
lauralorenz

[Failing test] gce-cos-master-scalability-100

Which jobs are failing?

release-master-blocking

gce-cos-master-scalability-100

Which tests are failing?

kubetest.ClusterLoaderV2 ci-kubernetes-e2e-gci-gce-scalability.Overall ClusterLoaderV2.load: [step: 19] Collecting pod startup latency ClusterLoaderV2.load overall (testing/load/config.yaml) (intermittently) ClusterLoaderV2.load: [step: 25] Waiting for 'delete objects' to be completed

Since when has it been failing?

3/15 6:47 PDT run; this run also happens to be the first at commit 1a5abe5d1

Testgrid link

https://testgrid.k8s.io/sig-release-master-blocking#gce-cos-master-scalability-100

Reason for failure (if possible)

[measurement call PodStartupLatency - PodStartupLatency error: pod startup: too high latency 99th percentile: got 10.515080076s expected: 5s]

Anything else we need to know?

No response

Relevant SIG(s)

/sig scalability

lauralorenz
lauralorenz

Am I crazy or is this robot not working?

/sig scalability

Previous