Tonic commited on
Commit
ad72fd3
1 Parent(s): ade11b4

improve interface

Browse files
Files changed (2) hide show
  1. app.py +17 -17
  2. globe.py +3 -3
app.py CHANGED
@@ -97,24 +97,24 @@ with gr.Blocks() as demo:
97
  value="You are a helpful AI assistant.",
98
  lines=2,
99
  placeholder="Set the AI's behavior and context..."
100
- )
101
- context = gr.Textbox(
102
- label="Context",
103
- lines=2,
104
- placeholder="Enter additional context information..."
105
- )
106
- max_tokens = gr.Slider(minimum=1, maximum=1024, value=256, step=1, label="Max Tokens")
107
- temperature = gr.Slider(minimum=0.1, maximum=2.0, value=0.7, step=0.1, label="Temperature")
108
- top_p = gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-p")
109
- use_pipeline = gr.Checkbox(label="Use Pipeline", value=False)
110
- use_tool = gr.Checkbox(label="Use Function Calling", value=False)
111
- with gr.Column(visible=False) as tool_options:
112
- tool_definition = gr.Code(
113
- label="Tool Definition (JSON)",
114
- value=customtool,
115
- lines=15,
116
- language="json"
117
  )
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
118
 
119
  def user(user_message, history):
120
  return "", history + [[user_message, None]]
 
97
  value="You are a helpful AI assistant.",
98
  lines=2,
99
  placeholder="Set the AI's behavior and context..."
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
100
  )
101
+ context = gr.Textbox(
102
+ label="Context",
103
+ lines=2,
104
+ placeholder="Enter additional context information..."
105
+ )
106
+ max_tokens = gr.Slider(minimum=1, maximum=1024, value=256, step=1, label="Max Tokens")
107
+ temperature = gr.Slider(minimum=0.1, maximum=2.0, value=0.7, step=0.1, label="Temperature")
108
+ top_p = gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-p")
109
+ use_pipeline = gr.Checkbox(label="Use Pipeline", value=False)
110
+ use_tool = gr.Checkbox(label="Use Function Calling", value=False)
111
+ with gr.Column(visible=False) as tool_options:
112
+ tool_definition = gr.Code(
113
+ label="Tool Definition (JSON)",
114
+ value=customtool,
115
+ lines=15,
116
+ language="json"
117
+ )
118
 
119
  def user(user_message, history):
120
  return "", history + [[user_message, None]]
globe.py CHANGED
@@ -14,14 +14,14 @@ presentation1 = """Try this model on [build.nvidia.com](https://build.nvidia.com
14
 
15
  **Model Dates:** 🤖Nemotron-Mini-4B-Instruct was trained between February 2024 and Aug 2024.
16
 
17
- ## License
18
 
19
  [NVIDIA Community Model License](https://huggingface.co/nvidia/Nemotron-Mini-4B-Instruct/blob/main/nvidia-community-model-license-aug2024.pdf)"""
20
 
21
  presentation2 = """
22
- ## Model Architecture
23
 
24
- Nemotron-Mini-4B-Instruct uses a model embedding size of 3072, 32 attention heads, and an MLP intermediate dimension of 9216. It also uses Grouped-Query Attention (GQA) and Rotary Position Embeddings (RoPE).
25
 
26
  **Architecture Type:** Transformer Decoder (auto-regressive language model)
27
 
 
14
 
15
  **Model Dates:** 🤖Nemotron-Mini-4B-Instruct was trained between February 2024 and Aug 2024.
16
 
17
+ ### License
18
 
19
  [NVIDIA Community Model License](https://huggingface.co/nvidia/Nemotron-Mini-4B-Instruct/blob/main/nvidia-community-model-license-aug2024.pdf)"""
20
 
21
  presentation2 = """
22
+ ### Model Architecture
23
 
24
+ 🤖Nemotron-Mini-4B-Instruct uses a model embedding size of 3072, 32 attention heads, and an MLP intermediate dimension of 9216. It also uses Grouped-Query Attention (GQA) and Rotary Position Embeddings (RoPE).
25
 
26
  **Architecture Type:** Transformer Decoder (auto-regressive language model)
27