Support to use torch.distributed.all_to_all_single to both dynamo and nondynamo case.
Note that there is a function signature mismatch between torch's all_to_all_single and xla op's AllToAll. To leverage the AllToAll op, we doesn't support specifying the input_split_sizes and output_split_sizes at this time. Check test_collective_ops_tpu.py for the usage.
Support to use torch.distributed.all_to_all_single to both dynamo and nondynamo case.
Note that there is a function signature mismatch between torch's all_to_all_single and xla op's AllToAll. To leverage the AllToAll op, we doesn't support specifying the
input_split_sizes
andoutput_split_sizes
at this time. Check test_collective_ops_tpu.py for the usage.