mirror of
https://github.com/huggingface/trl.git
synced 2025-10-21 11:33:51 +08:00
906 B
906 B
Models
With the AutoModelForCausalLMWithValueHead
class TRL supports all decoder model architectures in transformers such as GPT-2, OPT, and GPT-Neo. In addition, with AutoModelForSeq2SeqLMWithValueHead
you can use encoder-decoder architectures such as T5. TRL also requires reference models which are frozen copies of the model that is trained. With create_reference_model
you can easily create a frozen copy and also share layers between the two models to save memory.
PreTrainedModelWrapper
autodoc PreTrainedModelWrapper
AutoModelForCausalLMWithValueHead
autodoc AutoModelForCausalLMWithValueHead - init - forward - generate - _init_weights
AutoModelForSeq2SeqLMWithValueHead
autodoc AutoModelForSeq2SeqLMWithValueHead - init - forward - generate - _init_weights
create_reference_model
autodoc create_reference_model