microsoft / unilm

Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
https://aka.ms/GeneralAI
MIT License
19.09k stars 2.44k forks source link

Fine tuning LayoutLMV3 - YAML using layers 3, 5, 7, and 11 #1495

Open mapinusRND opened 2 months ago

mapinusRND commented 2 months ago

Describe Model I am using LayoutLM3:

You understood that the model is using layers 3, 5, 7, and 11, according to the cascade_layoutlmv3.yaml file.

However, in the original text, it states: "We adopt the standard practice to extract single-scale features from different Transformer layers, such as layers 4, 6, 8, and 12 of the LayoutLMv3 base model."

Is there a reason for changing the layers?