gamza commited on
Commit
4a7c6e4
โ€ข
1 Parent(s): 13b54a1

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +20 -34
app.py CHANGED
@@ -1,39 +1,25 @@
1
- from transformers import AutoModelForCausalLM, AutoTokenizer
2
  import gradio as gr
3
  import torch
4
-
5
-
6
- title = "๐Ÿค–AI ChatBot"
7
- description = "Building open-domain chatbots is a challenging area for machine learning research."
8
- examples = [["How are you?"]]
9
-
10
-
11
- tokenizer = AutoTokenizer.from_pretrained("microsoft/DialoGPT-large")
12
- model = AutoModelForCausalLM.from_pretrained("microsoft/DialoGPT-large")
13
-
14
-
15
- def predict(input, history=[]):
16
- # tokenize the new input sentence
17
- new_user_input_ids = tokenizer.encode(
18
- input + tokenizer.eos_token, return_tensors="pt"
19
- )
20
-
21
- # append the new user input tokens to the chat history
22
- bot_input_ids = torch.cat([torch.LongTensor(history), new_user_input_ids], dim=-1)
23
-
24
- # generate a response
25
- history = model.generate(
26
- bot_input_ids, max_length=4000, pad_token_id=tokenizer.eos_token_id
27
- ).tolist()
28
-
29
- # convert the tokens to text, and then split the responses into lines
30
- response = tokenizer.decode(history[0]).split("<|endoftext|>")
31
- # print('decoded_response-->>'+str(response))
32
- response = [
33
- (response[i], response[i + 1]) for i in range(0, len(response) - 1, 2)
34
- ] # convert to tuples of list
35
- # print('response-->>'+str(response))
36
- return response, history
37
 
38
 
39
  gr.Interface(
 
 
1
  import gradio as gr
2
  import torch
3
+ import pandas as pd
4
+ from sentence_transformers import SentenceTransformer
5
+ from sklearn.metrics.pairwise import cosine_similarity
6
+
7
+ title = "๐Ÿ€๊ณ ๋ฏผ ํ•ด๊ฒฐ ๋„์„œ ์ถ”์ฒœ ์ฑ—๋ด‡๐Ÿ€"
8
+ description = "๊ณ ๋ฏผ์ด ๋ฌด์—‡์ธ๊ฐ€์š”? ๊ณ ๋ฏผ ํ•ด๊ฒฐ์„ ๋„์™€์ค„ ์ฑ…์„ ์ถ”์ฒœํ•ด๋“œ๋ฆฝ๋‹ˆ๋‹ค"
9
+ examples = [["์š”์ฆ˜ ์ž ์ด ์•ˆ ์˜จ๋‹ค"]]
10
+
11
+
12
+ model = SentenceTransformer('jhgan/ko-sroberta-multitask')
13
+
14
+ def response(message):
15
+ embedding = model.encode(message)
16
+ df['distance'] = df['embedding'].map(lambda x: cosine_similarity([embedding], [x]).squeeze())
17
+ answer = df.loc[df['distance'].idxmax()]
18
+ Book_title = answer['์ œ๋ชฉ']
19
+ Book_author = answer['์ž‘๊ฐ€']
20
+ Book_publisher = answer['์ถœํŒ์‚ฌ']
21
+ Book_comment = answer['์„œํ‰']
22
+ return print(message)
 
 
 
 
 
 
 
 
 
 
 
 
 
23
 
24
 
25
  gr.Interface(