Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -1,4 +1,5 @@
|
|
1 |
import nltk
|
|
|
2 |
from nltk.stem.lancaster import LancasterStemmer
|
3 |
import numpy as np
|
4 |
import tflearn
|
@@ -7,6 +8,7 @@ import random
|
|
7 |
import json
|
8 |
import pickle
|
9 |
import gradio as gr
|
|
|
10 |
|
11 |
# Ensure necessary NLTK resources are downloaded
|
12 |
try:
|
@@ -48,7 +50,7 @@ except FileNotFoundError:
|
|
48 |
# Function to process user input into a bag-of-words format
|
49 |
def bag_of_words(s, words):
|
50 |
bag = [0 for _ in range(len(words))]
|
51 |
-
s_words = nltk.word_tokenize(s)
|
52 |
s_words = [stemmer.stem(word.lower()) for word in s_words if word.lower() in words]
|
53 |
for se in s_words:
|
54 |
for i, w in enumerate(words):
|
|
|
1 |
import nltk
|
2 |
+
nltk.download('punkt')
|
3 |
from nltk.stem.lancaster import LancasterStemmer
|
4 |
import numpy as np
|
5 |
import tflearn
|
|
|
8 |
import json
|
9 |
import pickle
|
10 |
import gradio as gr
|
11 |
+
from nltk.tokenize import word_tokenize
|
12 |
|
13 |
# Ensure necessary NLTK resources are downloaded
|
14 |
try:
|
|
|
50 |
# Function to process user input into a bag-of-words format
|
51 |
def bag_of_words(s, words):
|
52 |
bag = [0 for _ in range(len(words))]
|
53 |
+
s_words = word_tokenize(s) # Replaced nltk.word_tokenize(s)
|
54 |
s_words = [stemmer.stem(word.lower()) for word in s_words if word.lower() in words]
|
55 |
for se in s_words:
|
56 |
for i, w in enumerate(words):
|