Trained & Fine-Tuned via AlekseyKorshuk/gpt2-jokes Model On Hugging Face by Magical Macaronis
Trained using the Fraser & Jester dataset of an approximate 2 million reddit jokes. The primary intention of this product is to evoke hilarity, mainly to lighten someone’s mood, while fundamentally test the proficiency of AI in producing an emotion so simple and yet healthy through something as multi-faceted and variable as comedy.
Downloads
# Of Team Members
Hours Worked
# Of Approx Jokes Trained On
Along the road of getting the right model, there were many challenges. The first one was getting to understand the concepts behind and how to implement NLP models. Of course, over the course of the first and second weeks, we learned and overcame that obstacle. Next was finding the right model, as discussed in the rest of the slides. Everyone tried a different model, but eventually we landed upon a decent one. Finally, the quality of jokes is the last challenge. Filtering out explicitness (bad/inappropriate words) is being fixed through fine-tuning the model and using less explicit data. Time restrictions were also another factor for this project but if we were to continue more on this project, we would add multilingual support, interactive conversations, and integration with social media
The pre-trained model that we used INITIALLY was distil-GPT. This model was used for 99.5% test size, and used 0.5% trained data. Since this was 232,000 rows we used, the trained data was using 1160 rows. Initially, the validation rate was decreasing. Then errors started to pop up with the runtime, which made me change parameters in my program. Then the validation LOSS slightly fluctuated and started to increase after multiple epochs. Finally, with adjusting to A100 Gpu, it started to process the code faster.