Response Generation
Once ConversationModel is trained, the model can be loaded via from_pretrained.
Assume your trained model is saved under model directory. Then you can load it as follows.
>>> from convmodel import ConversationModel
>>> model = ConversationModel.from_pretrained("model")
After loading your model, generate method generates response from a given context.
All the options defined in a transformers generate method can be also available in the generate method.
One example below uses top_p and top_k options with do_sample.
>>> model.generate(context=["こんにちは"], do_sample=True, top_p=0.95, top_k=50)
ConversationModelOutput(responses=['こんにちは♪'], context=['こんにちは'])