| West et al.'s model from their "reflective decoding" paper. | |
| Sample usage: | |
| ```python | |
| import torch | |
| from modeling_opengpt2 import OpenGPT2LMHeadModel | |
| from padded_encoder import Encoder | |
| path_to_backward = 'danyaljj/opengpt2_pytorch_backward' | |
| encoder = Encoder() | |
| model_backward = OpenGPT2LMHeadModel.from_pretrained(path_to_backward) | |
| input = "until she finally won." | |
| input_ids = encoder.encode(input) | |
| input_ids = torch.tensor([input_ids[::-1] ], dtype=torch.int) | |
| print(input_ids) | |
| output = model_backward.generate(input_ids) | |
| output_text = encoder.decode(output.tolist()[0][::-1]) | |
| print(output_text) | |
| ``` | |
| Download the additional files from here: https://github.com/peterwestuw/GPT2ForwardBackward | |