Skip to content
Commit 8da9b1c5 authored by Liezl Puzon's avatar Liezl Puzon Committed by Facebook Github Bot
Browse files

Load a XLM model into transformer encoder / decoder for MT training (#629)

Summary:
Pull Request resolved: https://github.com/pytorch/fairseq/pull/629

Use GeLU as an alternate activation layer for ReLU.

Reviewed By: lematt1991

Differential Revision: D14689851

fbshipit-source-id: 7ec81fa34bc7bd0e1e43b337847ae932dcbf8b15
parent 8500bdd0
Loading
Loading
Loading
Loading
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment