Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

how to initialize n-gram tower and emb? #25

Open
FengYue95 opened this issue Jan 30, 2023 · 2 comments
Open

how to initialize n-gram tower and emb? #25

FengYue95 opened this issue Jan 30, 2023 · 2 comments

Comments

@FengYue95
Copy link

Hi~

1、Is ZEN trained from any base bert(e.g. google) or trained from scratch? If from scrach, I guess the n-gram emb is randomly initialized, If from base bert, the n-gram emb maybe the average of characters included?

2、According to "We use the same parameter setting for the n-gram encoder as in BERT" in the paper,I want to know that the params of n-gram encoder is shared and the same with bert tower(maybe the bottom six layer?),or is initialized and trained independently?

thank you~

@GuiminChen
Copy link
Collaborator

GuiminChen commented Jan 30, 2023 via email

@shizhediao
Copy link

shizhediao commented Sep 20, 2023

  1. There are two models in our paper. (R): randomly initialized parameters and (P): pre-trained model, which is the Google released Chinese BERT base model.
  2. Sorry I don't quite get your question, could you elaborate on it? Thanks

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants