bkoz commited on
Commit
eb72820
1 Parent(s): faee068
Files changed (1) hide show
  1. app.py +12 -12
app.py CHANGED
@@ -35,18 +35,18 @@ def load_model():
35
  ]
36
 
37
  for prompt in prompts:
38
- output = llm(
39
- prompt,
40
- max_tokens=512,
41
- temperature=0.4,
42
- grammar=grammar
43
- )
44
-
45
- s = output['choices'][0]['text']
46
- print(f'{s} , len(s) = {len(s)}')
47
- print(output['choices'])
48
- print(output['choices'][0]['text'])
49
- print()
50
 
51
 
52
  load_model()
 
35
  ]
36
 
37
  for prompt in prompts:
38
+ output = llm(
39
+ prompt,
40
+ max_tokens=512,
41
+ temperature=0.4,
42
+ grammar=grammar
43
+ )
44
+
45
+ s = output['choices'][0]['text']
46
+ print(f'{s} , len(s) = {len(s)}')
47
+ print(output['choices'])
48
+ print(output['choices'][0]['text'])
49
+ print()
50
 
51
 
52
  load_model()