Shortlink: goo.gl/wSuuS9
The github repository can be found at https://github.com/tensorflow/tensor2tensor/tree/master/tensor2tensor/data_generators/wikisum
Shortlink: goo.gl/wSuuS9
The github repository can be found at https://github.com/tensorflow/tensor2tensor/tree/master/tensor2tensor/data_generators/wikisum
Thanks for linking that, @peterjliu. Am I reading the README.md
correctly, here, that training uses a full transformer architecture, rather than a decoder-only architecture with memory-compressed attention?
TODO(rsepassi): Put actual results achieved on wikisum_web
and/or
wikisum_commoncrawl
and with what hparams_set
.
PROBLEM=wikisum_web # or wikisum_commoncrawl
t2t-trainer \
--problem=$PROBLEM \
--model=transformer \
--hparams_set=transformer_base \
--train_steps=250000 \
--eval_steps=100 \
--data_dir=$DATA_DIR \
--output_dir=$TRAIN_DIR
Does anyone have processed training examples (i.e., the output of step 3 here) available to share? I'm having trouble getting GCP to release IP addresses for data generation, so I'm hoping to be able to bypass this for the time being...
Also, as @nlothian and @hoang-ho have asked, are pre-trained model weights available anywhere?
Dear all,
Is there any available pre-trained model released for this wikisum problem? If there is, may I have the link to that pre-trained model?
Thank you so much