nugentc commited on
Commit
71f2227
1 Parent(s): 8ed8504

swap out grammar model

Browse files
Files changed (1) hide show
  1. app.py +11 -8
app.py CHANGED
@@ -1,8 +1,8 @@
1
- from transformers import AutoModelForCausalLM, AutoTokenizer, AutoModelForSeq2SeqLM
2
  tokenizer = AutoTokenizer.from_pretrained("microsoft/DialoGPT-large")
3
  model = AutoModelForCausalLM.from_pretrained("microsoft/DialoGPT-large")
4
- grammar_tokenizer = AutoTokenizer.from_pretrained("prithivida/grammar_error_correcter_v1")
5
- grammar_model = AutoModelForSeq2SeqLM.from_pretrained("prithivida/grammar_error_correcter_v1")
6
  import torch
7
  import gradio as gr
8
 
@@ -22,14 +22,17 @@ def chat(message, history):
22
 
23
 
24
  def feedback(text):
25
- tokenized_phrases = grammar_tokenizer([text], return_tensors='pt', padding=True)
26
- corrections = grammar_model.generate(**tokenized_phrases)
27
- corrections = grammar_tokenizer.batch_decode(corrections, skip_special_tokens=True)
 
 
28
  print("The corrections are: ", corrections)
29
- if corrections[0] == text:
30
  feedback = f'Looks good! Keep up the good work'
31
  else:
32
- feedback = f'\'{corrections[0]}\' might be a little better'
 
33
  return f'FEEDBACK: {feedback}'
34
 
35
  iface = gr.Interface(
 
1
+ from transformers import AutoModelForCausalLM, AutoTokenizer, AutoModelForSeq2SeqLM, T5ForConditionalGeneration, T5Tokenizer
2
  tokenizer = AutoTokenizer.from_pretrained("microsoft/DialoGPT-large")
3
  model = AutoModelForCausalLM.from_pretrained("microsoft/DialoGPT-large")
4
+ grammar_tokenizer = T5Tokenizer.from_pretrained('deep-learning-analytics/GrammarCorrector')
5
+ grammar_model = T5ForConditionalGeneration.from_pretrained('deep-learning-analytics/GrammarCorrector')
6
  import torch
7
  import gradio as gr
8
 
 
22
 
23
 
24
  def feedback(text):
25
+ # tokenized_phrases = grammar_tokenizer([text], return_tensors='pt', padding=True)
26
+ # corrections = grammar_model.generate(**tokenized_phrases)
27
+ # corrections = grammar_tokenizer.batch_decode(corrections, skip_special_tokens=True)
28
+ batch = grammar_tokenizer([input_text],truncation=True,padding='max_length',max_length=64, return_tensors="pt").to(torch_device)
29
+ corrections= grammar_model.generate(**batch,max_length=64,num_beams=2, num_return_sequences=num_return_sequences, temperature=1.5)
30
  print("The corrections are: ", corrections)
31
+ if len(corrections) == 0:
32
  feedback = f'Looks good! Keep up the good work'
33
  else:
34
+ suggestion = tokenizer.batch_decode(corrections[0], skip_special_tokens=True)
35
+ feedback = f'\'{suggestion}\' might be a little better'
36
  return f'FEEDBACK: {feedback}'
37
 
38
  iface = gr.Interface(