Shortlink: goo.gl/wSuuS9
The github repository can be found at https://github.com/tensorflow/tensor2tensor/tree/master/tensor2tensor/data_generators/wikisum
Shortlink: goo.gl/wSuuS9
The github repository can be found at https://github.com/tensorflow/tensor2tensor/tree/master/tensor2tensor/data_generators/wikisum
Does anyone have processed training examples (i.e., the output of step 3 here) available to share? I'm having trouble getting GCP to release IP addresses for data generation, so I'm hoping to be able to bypass this for the time being...
Also, as @nlothian and @hoang-ho have asked, are pre-trained model weights available anywhere?
Thanks for linking that, @peterjliu. Am I reading the
README.md
correctly, here, that training uses a full transformer architecture, rather than a decoder-only architecture with memory-compressed attention?Training
TODO(rsepassi): Put actual results achieved on
wikisum_web
and/orwikisum_commoncrawl
and with whathparams_set
.