Spaces:
Runtime error
Runtime error
File size: 2,236 Bytes
605bee0 183e4db 605bee0 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 |
from dataclasses import dataclass, field
from typing import Optional
@dataclass
class TrainArguments(object):
model_name_or_path: Optional[str] = field(
default="Open-Orca/OpenOrca-Platypus2-13B,
metadata={"help": "model_name_or_path"},
)
model_basename: Optional[str] = field(
default="gptq_model-4bit-128g",
metadata={"help": "model_basename"},
)
DEFAULT_SYSTEM_PROMPT: Optional[str] = field(
default="",
metadata={"help": "DEFAULT_SYSTEM_PROMPT"},
)
MAX_MAX_NEW_TOKENS: Optional[int] = field(
default=2048,
metadata={"help": "MAX_MAX_NEW_TOKENS"},
)
DEFAULT_MAX_NEW_TOKENS: Optional[int] = field(
default=1024,
metadata={"help": "MAX_MAX_NEW_TOKENS"},
)
MAX_INPUT_TOKEN_LENGTH: Optional[int] = field(
default=4000,
metadata={"help": "MAX_MAX_NEW_TOKENS"},
)
MODEL_PATH: Optional[str] = field(
default="TheBloke/Llama-2-7b-Chat-GPTQ",
metadata={"help": "MODEL_PATH"},
)
BACKEND_TYPE: Optional[str] = field(
default=".env.7b_gptq_example",
metadata={"help": "BACKEND_TYPE"},
)
LOAD_IN_8BIT: Optional[bool] = field(
default=True,
metadata={"help": "LOAD_IN_8BIT"},
)
LOAD_IN_4BIT: Optional[bool] = field(
default=False,
metadata={"help": "LOAD_IN_4BIT"},
)
DESCRIPTION: Optional[str] = field(
default="""
# llama2-webui
This is a chatbot based on Llama-2.
- Supporting models:
[Llama-2-7b](https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML)
[13b](https://huggingface.co/llamaste/Llama-2-13b-chat-hf)
[70b](https://huggingface.co/llamaste/Llama-2-70b-chat-hf)
all [Llama-2-GPTQ](https://huggingface.co/TheBloke/Llama-2-7b-Chat-GPTQ)
all [Llama-2-GGML](https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML)
- Supporting model backends:
[tranformers](https://github.com/huggingface/transformers)
[bitsandbytes(8-bit inference)](https://github.com/TimDettmers/bitsandbytes)
[AutoGPTQ(4-bit inference)](https://github.com/PanQiWei/AutoGPTQ)
[llama.cpp](https://github.com/ggerganov/llama.cpp)
""",
metadata={"help": "DESCRIPTION"},
) |