Also, remember to tag it as conversational in its Model Card (equivalently its README.md):. We will need this token when we build the Discord bot.įollow along with this section in my video to push the model. Obtain your API token by going to Edit profile > API Tokens. Sign up for Hugging Face and create a new model repository by clicking on New model. We will host the model on Hugging Face, which provides a free API for us to query the model. This isn't ideal as we want the conversation to be more organic. ![]() However, do take care not to overfit the model: If the model is trained for too many epochs, it may memorize the dataset and recite back lines from the dataset when we try to converse with it. The model will generally get smarter when it has more exposure to the dataset. This is the number of times that the model will cycle through the training dataset. You may also increase the number of training epochs by searching for num_train_epochs in the notebook. More parameters will allow the model to pick up more complexity from the dataset. Model size here refers to the number of parameters in the model. ![]() Want an even smarter and more eloquent model? Feel free to train a larger model like DialoGPT-medium or even DialoGPT-large. The model will be stored in a folder named output-small. I have about 700 lines and the training takes less than ten minutes. Running through the training section of the notebook should take less than half an hour.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |