Skip to content
This repository has been archived by the owner on Dec 16, 2022. It is now read-only.

Transformer toolkit updates #5270

Merged
merged 13 commits into from
Jun 21, 2021
Merged

Transformer toolkit updates #5270

merged 13 commits into from
Jun 21, 2021

Conversation

dirkgr
Copy link
Member

@dirkgr dirkgr commented Jun 17, 2021

This is spun out of some of the Tango stuff.

From the changelog:

  • The activation layer in the transformer toolkit now can be queried for its output dimension.
  • TransformerEmbeddings now takes, but ignores, a parameter for the attention mask. This is needed for compatibility with some other modules that get called the same way and use the mask.
  • TransformerPooler can now be instantiated from a pretrained transformer module, just like the other modules in the transformer toolkit.

@dirkgr dirkgr self-assigned this Jun 17, 2021
@dirkgr dirkgr marked this pull request as ready for review June 17, 2021 22:52
@dirkgr dirkgr requested a review from AkshitaB June 17, 2021 22:52
super().__init__(hidden_size, intermediate_size, activation, pool=True)

@classmethod
def _get_input_arguments(
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We don't require this method any longer. from_config takes care of what we need.

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Removed!

@dirkgr dirkgr enabled auto-merge (squash) June 21, 2021 17:14
@dirkgr dirkgr merged commit c8b8ed3 into main Jun 21, 2021
@dirkgr dirkgr deleted the TransformerToolkitUpdates branch June 21, 2021 17:28
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants