Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -59,9 +59,15 @@ def givetext(input_text,lmodel,ltokenizer):
|
|
59 |
print("BEFORE DOING TORCH.NO_GRAD()")
|
60 |
|
61 |
with torch.no_grad():
|
62 |
-
print("BEFORE RETURNING")
|
|
|
|
|
|
|
|
|
|
|
|
|
63 |
|
64 |
-
return (ltokenizer.decode(
|
65 |
#return (ltokenizer.decode(lmodel.generate(**model_input, max_new_tokens=1000)[0], skip_special_tokens=True))
|
66 |
#return (ltokenizer.decode(lmodel.generate(**model_input, max_new_tokens=100)[0], skip_special_tokens=True))
|
67 |
except Exception as error:
|
|
|
59 |
print("BEFORE DOING TORCH.NO_GRAD()")
|
60 |
|
61 |
with torch.no_grad():
|
62 |
+
#print("BEFORE RETURNING")
|
63 |
+
|
64 |
+
print("BEFORE GENERATING LMODEL")
|
65 |
+
|
66 |
+
lmodel_generated = lmodel.generate(**model_input, max_new_tokens=1000, device="cpu")[0]
|
67 |
+
|
68 |
+
print("BEFORE GENERATING LTOKENIZER")
|
69 |
|
70 |
+
return (ltokenizer.decode(lmodel_generated, skip_special_tokens=True))
|
71 |
#return (ltokenizer.decode(lmodel.generate(**model_input, max_new_tokens=1000)[0], skip_special_tokens=True))
|
72 |
#return (ltokenizer.decode(lmodel.generate(**model_input, max_new_tokens=100)[0], skip_special_tokens=True))
|
73 |
except Exception as error:
|