Thalirajesh's picture
Update README.md
025b14e verified
---
language: en
widget:
- text: "Arts"
example_title: "Arts"
- text: "Space"
example_title: "Space"
- text: "Time"
example_title: "Time"
tags:
- text-generation
- creative-writing
- essay-writing
inference:
parameters:
max_length: 400
num_beams: 10
early_stopping: true
temperature: 0.3
no_repeat_ngram_size: 2
num_return_sequences: 2
---
Introduction:
This repository contains a finetuned DistilGPT2 model for generating diverse essays on topics spanning Arts, Science, and Culture.
Dataset:
The training dataset comprises 2000+ essays covering diverse topics in Arts, Science, and Culture. These essays are written by human experts and contain a diverse set of opinions and knowledge, ensuring that the model learns from high-quality and diverse content.
Model Training:
- epoch: 50
- training_loss: 2.473200
- validation_loss: 4.569556
- perplexities: [517.4149169921875, 924.535888671875, 704.73291015625, 465.9677429199219, 577.629150390625, 443.994140625, 770.1861572265625, 683.028076171875, 1017.7510375976562, 880.795166015625]
- mean_perplexity: 698.603519
Description:
The model achieved a mean perplexity of 698.603519 on the validation set, indicating its ability to generate diverse and high-quality essays on the given topics.
During Text Generation, the following parameters are used:
- `max_length`: The maximum length of the generated text, set to 400 tokens.
- `num_beams`: The number of beams for beam search, set to 10. A higher value will increase the diversity of the generated text but may also increase the inference time.
- `early_stopping`: If set to True, the generation will stop as soon as the end-of-sequence token is generated.
- `temperature`: The sampling temperature, is set to 0.3.
- `no_repeat_ngram_size`: The size of the n-gram window to avoid repetitions, set to 2.
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64fec5de57ccb8f1bdfbec54/ac89INQ8czj1u6WApI20J.png)
Find the kaggle notebook for this project at
[Kaggle Notebook](https://www.kaggle.com/code/vignesharjunraj/finetuned-distilgpt2-llm-for-essays-400-words/)