We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
bert的中文vocab.txt的汉字太少了,请问垂直领域遇到这样情况是怎么增加自己的token呢, 【unused】不够用
The text was updated successfully, but these errors were encountered:
1、把它们加入到vocab.txt里边 2、通过compound_tokens参数追加。
compound_tokens
以上只适合bert4keras,可以仔细琢磨一下训练脚本中追加词的方法:https://github.com/ZhuiyiTechnology/WoBERT/blob/master/train.py
Sorry, something went wrong.
您好,我无法下载文件请问还有没有别的办法下载模型
已恢复下载
No branches or pull requests
bert的中文vocab.txt的汉字太少了,请问垂直领域遇到这样情况是怎么增加自己的token呢, 【unused】不够用
The text was updated successfully, but these errors were encountered: