Open sxjscience opened 3 years ago
Some items:
On the Text Prediction - Part 1: Quickstart of Pretrained Backbones
we will load two datasets using the nlp_data command...
but the immediate next block are imports. Consider mentioning that the first block is import.Let’s download two datasets from the GLUE benchmark: - The Standford Sentiment Treebank (SST-2) - Semantic Textual Similarity Benchmark (STS-B)
A bunch of recent papers, especially BERT, have led a new trend for solving NLP problems: - Pretrain a backbone model on a large corpus, - Finetune the backbone to solve the specific NLP task.
According to the paper and our own experiments, MobileBERT performs similar to BERT-base...
, while useful as an introduction to the next tutorial, diverges from the topic of quick start. Consider concluding with something like "In this tutorial, we learned how to use pre-trained backbone networks to quickly start ... Next, we will ..."On the Text Prediction - Part2: MobileBERT for Text Prediction
Now you have learned 1) the basics about Gluon, 2) how to ...
). Instead we could adjust the wording as "In Part 1 Quickstart (link), we learned ..."On the Question Answering with GluonNLP
https://github.com/dmlc/gluon-nlp/tree/master/scripts/question_answering
should link to the example page so that we always refer to the source code in the same version.!wget -O google_electra_base_squad2.0_8160.params https://gluon-nlp-log.s3.amazonaws.com/squad_training_log/fintune_google_electra_base_squad_2.0/google_electra_base_squad2.0_8160.params
On the Tokenization - Part1: Basic Usage of Tokenizer and Vocabulary
text processing workflow: raw text => normalized (cleaned) text => tokens => network
can be replaced with a diagram in which examples of each step can be shown.On the Tokenization - Part2: Learn Subword Models with GluonNLP
On the Tokenization - Part3: Download Data from Wikipedia and Learn Subword
For the tokenization notebooks, one pressing need I see is that there should be a reference page for the functionality of the CLIs in the API section. Otherwise, short of reading the code, it's hard for users to discover the features in them.
On Compile NLP Models - Convert GluonNLP Models to TVM
compile_tvm_graph_runtime
is very long and there are opportunities for breaking up the logic into smaller functions. For example, graph construction can happen separately from compilation.
Description
https://github.com/dmlc/gluon-nlp/pull/1374 has been merged so we have fixed the warnings in our documents. However, the current structure of the website is not very satisfactory and we should try to improve the layout and also add more tutorials.
Help needed here.
References
@dmlc/gluon-nlp-committers @Cli212 @yongyi-wu @xinyual @barry-jin