Spaces:
Runtime error
Runtime error
updated logging
Browse files
app.py
CHANGED
|
@@ -30,13 +30,16 @@ def user(message, history):
|
|
| 30 |
|
| 31 |
|
| 32 |
def chat(history, top_p, top_k, temperature):
|
|
|
|
|
|
|
| 33 |
# Initialize a StopOnTokens object
|
| 34 |
stop = StopOnTokens()
|
| 35 |
|
| 36 |
# Construct the input message string for the model by concatenating the current system message and conversation history
|
| 37 |
messages = "".join(["".join(["\n<human>:"+item[0], "\n<bot>:"+item[1]]) #curr_system_message +
|
| 38 |
for item in history])
|
| 39 |
-
|
|
|
|
| 40 |
# Tokenize the messages string
|
| 41 |
model_inputs = tok([messages], return_tensors="pt").to("cuda")
|
| 42 |
streamer = TextIteratorStreamer(
|
|
|
|
| 30 |
|
| 31 |
|
| 32 |
def chat(history, top_p, top_k, temperature):
|
| 33 |
+
|
| 34 |
+
print(f"history is - {history}")
|
| 35 |
# Initialize a StopOnTokens object
|
| 36 |
stop = StopOnTokens()
|
| 37 |
|
| 38 |
# Construct the input message string for the model by concatenating the current system message and conversation history
|
| 39 |
messages = "".join(["".join(["\n<human>:"+item[0], "\n<bot>:"+item[1]]) #curr_system_message +
|
| 40 |
for item in history])
|
| 41 |
+
print(f"messages is - {messages}")
|
| 42 |
+
|
| 43 |
# Tokenize the messages string
|
| 44 |
model_inputs = tok([messages], return_tensors="pt").to("cuda")
|
| 45 |
streamer = TextIteratorStreamer(
|