use r22.12
set
model_transaction_policy {
decoupled: True
}
dynamic_batching {
max_queue_delay_microseconds: 5000000
}
send two requests get max_batchs_size=1, The next request must wait for the previous request to end before starting
but set decoupled : False
send two requests get max_batchs_size=2, two requests returned simultaneously
Description
Reproduced Steps