Unverified Commit 4d1a3ffd authored by Patrick von Platen's avatar Patrick von Platen Committed by GitHub
Browse files

[EncoderDecoder] Add xlm-roberta to encoder decoder (#6878)

* finish xlm-roberta

* finish docs

* expose XLMRobertaForCausalLM
parent 31199263
...@@ -56,6 +56,13 @@ XLMRobertaModel ...@@ -56,6 +56,13 @@ XLMRobertaModel
:members: :members:
XLMRobertaForCausalLM
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
.. autoclass:: transformers.XLMRobertaForCausalLM
:members:
XLMRobertaForMaskedLM XLMRobertaForMaskedLM
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
...@@ -130,4 +137,4 @@ TFXLMRobertaForQuestionAnswering ...@@ -130,4 +137,4 @@ TFXLMRobertaForQuestionAnswering
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
.. autoclass:: transformers.TFXLMRobertaForQuestionAnswering .. autoclass:: transformers.TFXLMRobertaForQuestionAnswering
:members: :members:
\ No newline at end of file
...@@ -418,6 +418,7 @@ if is_torch_available(): ...@@ -418,6 +418,7 @@ if is_torch_available():
) )
from .modeling_xlm_roberta import ( from .modeling_xlm_roberta import (
XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST,
XLMRobertaForCausalLM,
XLMRobertaForMaskedLM, XLMRobertaForMaskedLM,
XLMRobertaForMultipleChoice, XLMRobertaForMultipleChoice,
XLMRobertaForQuestionAnswering, XLMRobertaForQuestionAnswering,
......
...@@ -156,6 +156,7 @@ from .modeling_xlm import ( ...@@ -156,6 +156,7 @@ from .modeling_xlm import (
XLMWithLMHeadModel, XLMWithLMHeadModel,
) )
from .modeling_xlm_roberta import ( from .modeling_xlm_roberta import (
XLMRobertaForCausalLM,
XLMRobertaForMaskedLM, XLMRobertaForMaskedLM,
XLMRobertaForMultipleChoice, XLMRobertaForMultipleChoice,
XLMRobertaForQuestionAnswering, XLMRobertaForQuestionAnswering,
...@@ -255,6 +256,7 @@ MODEL_WITH_LM_HEAD_MAPPING = OrderedDict( ...@@ -255,6 +256,7 @@ MODEL_WITH_LM_HEAD_MAPPING = OrderedDict(
MODEL_FOR_CAUSAL_LM_MAPPING = OrderedDict( MODEL_FOR_CAUSAL_LM_MAPPING = OrderedDict(
[ [
(CamembertConfig, CamembertForCausalLM), (CamembertConfig, CamembertForCausalLM),
(XLMRobertaConfig, XLMRobertaForCausalLM),
(RobertaConfig, RobertaForCausalLM), (RobertaConfig, RobertaForCausalLM),
(BertConfig, BertLMHeadModel), (BertConfig, BertLMHeadModel),
(OpenAIGPTConfig, OpenAIGPTLMHeadModel), (OpenAIGPTConfig, OpenAIGPTLMHeadModel),
......
...@@ -18,6 +18,7 @@ ...@@ -18,6 +18,7 @@
from .configuration_xlm_roberta import XLMRobertaConfig from .configuration_xlm_roberta import XLMRobertaConfig
from .file_utils import add_start_docstrings from .file_utils import add_start_docstrings
from .modeling_roberta import ( from .modeling_roberta import (
RobertaForCausalLM,
RobertaForMaskedLM, RobertaForMaskedLM,
RobertaForMultipleChoice, RobertaForMultipleChoice,
RobertaForQuestionAnswering, RobertaForQuestionAnswering,
...@@ -67,6 +68,19 @@ class XLMRobertaModel(RobertaModel): ...@@ -67,6 +68,19 @@ class XLMRobertaModel(RobertaModel):
config_class = XLMRobertaConfig config_class = XLMRobertaConfig
@add_start_docstrings(
"XLM-RoBERTa Model with a `language modeling` head on top for CLM fine-tuning.",
XLM_ROBERTA_START_DOCSTRING,
)
class XLMRobertaForCausalLM(RobertaForCausalLM):
"""
This class overrides :class:`~transformers.RobertaForCausalLM`. Please check the
superclass for the appropriate documentation alongside usage examples.
"""
config_class = XLMRobertaConfig
@add_start_docstrings( @add_start_docstrings(
"""XLM-RoBERTa Model with a `language modeling` head on top. """, """XLM-RoBERTa Model with a `language modeling` head on top. """,
XLM_ROBERTA_START_DOCSTRING, XLM_ROBERTA_START_DOCSTRING,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment