Here, the API responds with DEADLINE_EXCEEDED, which may mean it's a real problem. Need to reproduce locally, then check if increasing the timeout in gRPC service config solves it, and if it does, make the change internally to increase the timeout.
I'll it.skip the test for now.
(node:343) UnhandledPromiseRejectionWarning: Error: 4 DEADLINE_EXCEEDED: Deadline exceeded
at Object.callErrorFromStatus (/workspace/node_modules/@[grpc/grpc-js/build/src/call.js:31](https://cs.corp.google.com/piper///depot/google3/grpc/grpc-js/build/src/call.js?l=31):19)
at Object.onReceiveStatus (/workspace/node_modules/@[grpc/grpc-js/build/src/client.js:190](https://cs.corp.google.com/piper///depot/google3/grpc/grpc-js/build/src/client.js?l=190):52)
at Object.onReceiveStatus (/workspace/node_modules/@[grpc/grpc-js/build/src/client-interceptors.js:365](https://cs.corp.google.com/piper///depot/google3/grpc/grpc-js/build/src/client-interceptors.js?l=365):141)
at Object.onReceiveStatus (/workspace/node_modules/@[grpc/grpc-js/build/src/client-interceptors.js:328](https://cs.corp.google.com/piper///depot/google3/grpc/grpc-js/build/src/client-interceptors.js?l=328):181)
at /workspace/node_modules/@[grpc/grpc-js/build/src/call-stream.js:188](https://cs.corp.google.com/piper///depot/google3/grpc/grpc-js/build/src/call-stream.js?l=188):78
at processTicksAndRejections ([internal/process/task_queues.js:79](https://cs.corp.google.com/piper///depot/google3/internal/process/task_queues.js?l=79):11)
for call at
at ServiceClientImpl.makeUnaryRequest (/workspace/node_modules/@[grpc/grpc-js/build/src/client.js:160](https://cs.corp.google.com/piper///depot/google3/grpc/grpc-js/build/src/client.js?l=160):30)
at ServiceClientImpl.<anonymous> (/workspace/node_modules/@[grpc/grpc-js/build/src/make-client.js:105](https://cs.corp.google.com/piper///depot/google3/grpc/grpc-js/build/src/make-client.js?l=105):19)
at /[workspace/build/src/v1/data_catalog_client.js:233](https://cs.corp.google.com/piper///depot/google3/workspace/build/src/v1/data_catalog_client.js?l=233):29
at wrappedCall (/[workspace/node_modules/google-gax/build/src/paginationCalls/pagedApiCaller.js:86](https://cs.corp.google.com/piper///depot/google3/workspace/node_modules/google-gax/build/src/paginationCalls/pagedApiCaller.js?l=86):20)
at /[workspace/node_modules/google-gax/build/src/normalCalls/timeout.js:44](https://cs.corp.google.com/piper///depot/google3/workspace/node_modules/google-gax/build/src/normalCalls/timeout.js?l=44):16
at repeat (/[workspace/node_modules/google-gax/build/src/normalCalls/retries.js:80](https://cs.corp.google.com/piper///depot/google3/workspace/node_modules/google-gax/build/src/normalCalls/retries.js?l=80):25)
at Immediate._onImmediate (/[workspace/node_modules/google-gax/build/src/normalCalls/retries.js:118](https://cs.corp.google.com/piper///depot/google3/workspace/node_modules/google-gax/build/src/normalCalls/retries.js?l=118):13)
at processImmediate ([internal/timers.js:463](https://cs.corp.google.com/piper///depot/google3/internal/timers.js?l=463):21)
(node:343) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). To terminate the node process on unhandled promise rejection, use the CLI flag `--unhandled-rejections=strict` (see [https://nodejs.org/api/cli.html#cli_unhandled_rejections_mode](https://www.google.com/url?q=https://nodejs.org/api/cli.html%23cli_unhandled_rejections_mode&sa=D)). (rejection id: 1)
(node:343) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code.
1) should search data assets in project
Log: https://fusion2.corp.google.com/invocations/5bddf53d-130c-4b57-8f1e-66a5e83584b0/targets/cloud-devrel%2Fclient-libraries%2Fnodejs%2Fpresubmit%2Fgoogleapis%2Fnodejs-datacatalog%2Fnode12%2Fsamples-test/log
Here, the API responds with
DEADLINE_EXCEEDED
, which may mean it's a real problem. Need to reproduce locally, then check if increasing the timeout in gRPC service config solves it, and if it does, make the change internally to increase the timeout.I'll
it.skip
the test for now.