Last active
August 11, 2019 13:39
-
-
Save bfarzin/f9407c0b0f2690f36dd6d51c8ef56944 to your computer and use it in GitHub Desktop.
i believe this is how i should use it:
mycust_tok = CustomTokenizer(SPTokenizer,model_prefix)
But i still have the error. Can you help with the full modified code?
itos
was wrong also. I updated the example above.
Thank you so much. Sorry I didnt notice the change earlier. It works. Much appreciated.
No problem. I am glad I cleaned it up for my own good!
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I am not too sure. I have moved past this example and now have better code for the custom tokenizer. (which allows it to be saved and applies
EncodeAsPieces
which will return components of the sub-word rather than the ID (the numericaliztion!) See if this help at all or if you get the same errors: