mlcommons / training_policies

Issues related to MLPerf™ training policies, including rules and suggested changes
https://mlcommons.org/en/groups/training
Apache License 2.0
93 stars 66 forks source link

Update training_rules.adoc #448

Open mrinal-gc opened 3 years ago

mrinal-gc commented 3 years ago

packing rule update

github-actions[bot] commented 3 years ago

MLCommons CLA bot All contributors have signed the MLCommons CLA ✍️ ✅

guschmue commented 3 years ago

recheck

guschmue commented 3 years ago

recheck

mrinal-gc commented 3 years ago

@johntran-nv @petermattson Could you review this?

petermattson commented 3 years ago

Closer! :-)

IMO, the only change needed is to add this paragraph:

(Un)padding or (un)packing are both allowed as offline or online preprocessing steps, including removal or addition of zero tokens. When packing, It is permitted to reorder and compress the dataset. However, the overall data traversal order, taking into account any packing, must still be as a random as the reference application. For instance: It is allowed to (a) pack items into groups offline then to randomly reorder the groups each run or to (b) randomly order the items then pack them into groups as traversed online provided that in both cases the groups are much smaller than the overall dataset. It is not allowed to sort for packing and use the same sorted order for every run.

I'd revert the changes and stick this on the end of the first CLOSED: para in the section. WDYT?

mrinal-gc commented 3 years ago

Way more elegant! I’ll send an update

Get Outlook for iOShttps://aka.ms/o0ukef


From: Peter Mattson @.> Sent: Thursday, April 29, 2021 6:53:58 PM To: mlcommons/training_policies @.> Cc: Mrinal Iyer @.>; Mention @.> Subject: Re: [mlcommons/training_policies] Update training_rules.adoc (#448)

Closer! :-)

IMO, the only change needed is to add this paragraph:

(Un)padding or (un)packing are both allowed as offline or online preprocessing steps, including removal or addition of zero tokens. When packing, It is permitted to reorder and compress the dataset. However, the overall data traversal order, taking into account any packing, must still be as a random as the reference application. For instance: It is allowed to (a) pack items into groups offline then to randomly reorder the groups each run or to (b) randomly order the items then pack them into groups as traversed online provided that in both cases the groups are much smaller than the overall dataset. It is not allowed to sort for packing and use the same sorted order for every run.

I'd revert the changes and stick this on the end of the first CLOSED: para in the section. WDYT?

— You are receiving this because you were mentioned. Reply to this email directly, view it on GitHubhttps://github.com/mlcommons/training_policies/pull/448#issuecomment-829741635, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AT4F5X3SNNAKF6YFT5KWY63TLIETNANCNFSM432V5XAQ.

We have updated our privacy policy, which contains important information about how we collect and process your personal data. To read the policy, please click herehttp://www.graphcore.ai/privacy

This email and its attachments are intended solely for the addressed recipients and may contain confidential or legally privileged information. If you are not the intended recipient you must not copy, distribute or disseminate this email in any way; to do so may be unlawful.

Any personal data/special category personal data herein are processed in accordance with UK data protection legislation. All associated feasible security measures are in place. Further details are available from the Privacy Notice on the website and/or from the Company.

Graphcore Limited (registered in England and Wales with registration number 10185006) is registered at 107 Cheapside, London, UK, EC2V 6DN. This message was scanned for viruses upon transmission. However Graphcore accepts no liability for any such transmission.

mrinal-gc commented 3 years ago

@petermattson done! Thanks