Transformers
Safetensors
English
medical
text-generation-inference
llm
biomistral
Inference Endpoints
siyah1's picture
Update README.md
85fd568 verified
metadata
library_name: transformers
tags:
  - medical
  - text-generation-inference
  - llm
  - biomistral
license: apache-2.0
datasets:
  - ruslanmv/ai-medical-chatbot
  - lavita/ChatDoctor-HealthCareMagic-100k
language:
  - en
metrics:
  - rouge
  - bleu

Model Card for BioMistral Multi-Turn Doctor Conversation Model

Model Details

Model Description

This model is a fine-tuned version of the BioMistral model, specifically tailored for multi-turn doctor-patient conversations. It leverages the powerful language generation capabilities of BioMistral to provide accurate and context-aware responses in medical dialogue scenarios.

  • Developed by: Siyahul Haque T P
  • Model type: Text-generation (LLM)
  • Language(s) (NLP): English (en)
  • License: Apache-2.0
  • Finetuned from model: BioMistral

Uses

Direct Use

This model can be directly used for generating responses in multi-turn medical conversations, making it useful for applications like virtual health assistants and medical chatbots.

Downstream Use

This model can be further fine-tuned or integrated into larger healthcare applications, such as patient management systems or automated symptom checkers.

Out-of-Scope Use

The model is not suitable for use in emergency medical situations, providing final diagnoses, or replacing professional medical advice.

Bias, Risks, and Limitations

The model may reflect biases present in the training data, including underrepresentation of certain medical conditions or demographic groups. The model should not be used as a sole source of medical information and must be supervised by qualified healthcare professionals.

Recommendations

Users should be aware of the potential biases and limitations of the model. It is recommended to use the model as a supplementary tool rather than a primary source of medical advice.

How to Get Started with the Model

from transformers import AutoModelForCausalLM, AutoTokenizer

# Load the tokenizer and model from the Hugging Face Hub
tokenizer = AutoTokenizer.from_pretrained("BioMistral/BioMistral-7B")
model = AutoModelForCausalLM.from_pretrained("siyah1/BioMistral-7b-Chat-Doctor")

# Example input: patient describing a symptom
input_text = "Hello, doctor, I have a headache."

# Tokenize the input text
inputs = tokenizer(input_text, return_tensors="pt")

# Generate a response from the model
outputs = model.generate(**inputs, max_length=100, num_return_sequences=1)

# Decode the generated response
response = tokenizer.decode(outputs[0], skip_special_tokens=True)

# Print the model's response
print("Doctor:", response)