Open christ1ne opened 4 years ago
I think the rules allow unpadding.
Section 7.1 says, "May pad to arbitrary size (don’t be creative)." I interpret that to allow unpadded inputs.
Wait is this a question about training or inference? If this is a training question, I retract my previous answer. If this is an inference question, it is on the wrong repository.
I observed at MLPerf training first. We should clarify on inference too.
SWG:
We agree we think this is allowed but it looks like this isn't clear in the rules, we should clarify the rules here.
We should probably add wording under "9.5. Equivalence exceptions" to capture this issue.
We will revisit next week with wording to capture this issue.
SWG:
Link to Proposal: https://docs.google.com/document/d/12grytcLPkQhU12pR0O02YtSiB87ezrVlNr9nrywCGtI/edit#
SWG:
In summary, this proposal is pack instead of pad examples for the purposes of not processing padding and to better load balance examples across devices. There are details about how to keep mathematical equivalence, including unpacking for certain operations/layers.
Things we can clarify in the rules:
Re-write the pre-processing rules to read as "follow the reference with these exceptions"
AI(NV) look at how to craft easier to understand rules here.
The reference has padded to a fixed sequence length. Does submitted code need to do that or we can allow unpadding?