Edit model card

Model Summary

Qwen2-7B-Instruct-Better-Translation is a fine-tuned language model based on Qwen2-7B-Instruct, specifically optimized for improving English-to-Chinese translation. The model was fine-tuned using Direct Preference Optimization (DPO) with a custom dataset that prioritizes fluent, idiomatic translations (chosen) over literal translations (rejected).

Developers: sevenone

  • License: Qwen2 License
  • Base Model: Qwen2-7B-Instruct
  • Model Size: 7B
  • Context Length: 131,072 tokens (inherits from Qwen2-7B-Instruct)

For more details, please refer to our GitHub.

1. Introduction

Qwen2-7B-Instruct-Better-Translation is designed to provide high-quality English-to-Chinese translations, particularly focusing on producing natural, idiomatic translations instead of literal, word-for-word translations. The fine-tuning process involved using a preference dataset where the chosen translations were idiomatic and the rejected translations were more literal. This model is ideal for users who need accurate and fluent translations for complex or nuanced English text.

2. Training Details

The model was fine-tuned using Direct Preference Optimization (DPO), a method that optimizes the model to prefer certain outputs over others based on user-provided preferences. The training dataset consisted of English source sentences, with corresponding translations labeled as either "chosen" (idiomatic) or "rejected" (literal).

  • Training Framework: Hugging Face Transformers
  • Optimizer: AdamW
  • Training Method: Lora with direct preference optimization
  • Training Data: Custom preference dataset for English-to-Chinese translation
  • Preference Type: Favoring idiomatic translations (chosen) over literal translations (rejected)

3. Requirements

To use this model, please ensure you have installed transformers>=4.37.0 to avoid any compatibility issues.

4. Usage

You can load and use the model to translate English to Chinese as shown in the following code snippet:

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model_id = "sevenone/Qwen2-7B-Instruct-Better-Translation"
device = "cuda"  # load onto GPU if available

tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
    model_id,
    torch_dtype="auto",
    device_map="auto"
)

prompt = "Translate the following sentence to Chinese: 'Artificial intelligence is transforming industries worldwide.'"
messages = [
    {"role": "system", "content": "You are a helpful assistant."},
    {"role": "user", "content": prompt}
]

# Apply the chat template for better generation
text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True
)
model_inputs = tokenizer([text], return_tensors="pt").to(device)

# Generate translation
generated_ids = model.generate(
    model_inputs.input_ids,
    max_new_tokens=512
)
generated_ids = [
    output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]

response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
print(response)

5. Citation

If sevenone/qwen2-7b-instruct-better-translation is helpful in your work, please kindly cite as:

@misc{sevenone_2024,
    author       = {sevenone},
    title        = {Qwen2-7B-Instruct-Better-Translation},
    year         = 2024,
    url          = {https://huggingface.co/sevenone/Qwen2-7B-Instruct-Better-Translation},
    publisher    = {Hugging Face}
}
Downloads last month
10
Safetensors
Model size
7.62B params
Tensor type
BF16
·
Inference Examples
Unable to determine this model's library. Check the docs .

Model tree for sevenone/Qwen2-7B-Instruct-Better-Translation

Base model

Qwen/Qwen2-7B
Finetuned
(30)
this model