|
--- |
|
datasets: |
|
- bitext/Bitext-retail-banking-llm-chatbot-training-dataset |
|
language: |
|
- en |
|
base_model: |
|
- google-t5/t5-small |
|
pipeline_tag: question-answering |
|
tags: |
|
- fintech |
|
- retail-banking |
|
- fine-tuning |
|
- chatbot |
|
- llm |
|
license: cdla-sharing-1.0 |
|
--- |
|
# fintech-chatbot-t5 |
|
|
|
## Model Description |
|
This model was fine-tuned using a [retail banking chatbot dataset](https://huggingface.co/datasets/bitext/Bitext-retail-banking-llm-chatbot-training-dataset/tree/main). It is based on the T5-small architecture and is capable of answering common banking-related queries like account balances, transaction details, card activations, and more. |
|
|
|
The model has been trained to generate responses to banking-related customer queries and is suited for use in automated customer service systems or virtual assistants. |
|
|
|
## Model Details |
|
- **Model Type:** T5-small |
|
- **Training Dataset:** [retail banking chatbot dataset](https://huggingface.co/datasets/bitext/Bitext-retail-banking-llm-chatbot-training-dataset/tree/main) |
|
- **Tasks:** Natural Language Generation (NLG) |
|
- **Languages Supported:** English |
|
|
|
## Training Details |
|
- **Number of Epochs:** 3 |
|
- **Training Loss:** 0.79 |
|
- **Evaluation Loss:** 0.46 |
|
- **Evaluation Metric:** Mean Squared Error |
|
- **Batch Size:** 8 |
|
- |
|
|
|
## How to Use the Model |
|
You can load and use this model with the following code: |
|
|
|
```python |
|
from transformers import T5Tokenizer, T5ForConditionalGeneration |
|
|
|
tokenizer = T5Tokenizer.from_pretrained("cuneytkaya/fintech-chatbot-t5") |
|
model = T5ForConditionalGeneration.from_pretrained("cuneytkaya/fintech-chatbot-t5") |
|
|
|
input_text = "How can I activate my credit card?" |
|
inputs = tokenizer.encode(input_text, return_tensors="pt") |
|
outputs = model.generate(inputs) |
|
|
|
print(tokenizer.decode(outputs[0])) |
|
|
|
|
|
|