Edit model card

murakami_rugpt3small

Model description

Fine-tuned from sberbank-ai/rugpt3small_based_on_gpt2

Intended uses & limitations

Generate novels in the style of Haruki Murakami, in Russian.

Just my exercise to learn the Huggingface ecosystem, really.

Training and evaluation data

Fine-tuned on murakami dataset, which was built from Russian translations of novels by Haruki Murakami.

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • gradient_checkpointing: True
  • num_epochs: 3.0
  • mixed_precision_training: Native AMP (fp16=True)

Framework versions

  • Transformers 4.27.3
  • Pytorch 2.0.0+cu117
  • Datasets 2.10.1
  • Tokenizers 0.13.2
Downloads last month
11
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.