Skip to content

mlp in language encoder #35

@zwx8981

Description

@zwx8981

embedded = self.mlp(embedded) # (n, seq_len, word_vec_size)

Hi, I find a mlp in the language encoder, which is used to project the original word embeddings into new representations. While it is not mentioned in the paper, so I am just curious about why you do this in your actual implementation? Is it a necessary operation to make the model work? How would it affect the model's performance?

Looking forward to hearing from you. Thank you so much.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions