导入数据
with open(intents.json) as file:
data = json.load(file)
try:
with open(data.pickle, rb) as f:
words, labels, training, output = pickle.load(f)
except:
words = []
labels = []
docs_x = []
docs_y = []
数据预处理
for intent in data[intents]:
for pattern in intent[patterns]:
wrds = nltk.word_tokenize(pattern)
words.extend(wrds)
docs_x.append(wrds)
docs_y.append(intent[tag])
if intent[tag] not in labels:
labels.append(intent[tag])
words = [stemmer.stem(w.lower()) for w in words if w != ?]
words = sorted(list(set(words)))
labels = sorted(labels)
training = []
output = []
out_empty = [0 for _ in range(len(labels))]
for x, doc in enumerate(docs_x):
bag = []
wrds = [stemmer.stem(w.lower()) for w in doc]
for w in words:
if w in wrds:
bag.append(1)
else:
bag.append(0)
training = numpy.array(training)
output = numpy.array(output)
with open(data.pickle, wb) as f:
pickle.dump((words, labels, training, output), f)
tensorflow.reset_default_graph()
net = tflearn.input_data(shape=[None, len(training[0])])
net = tflearn.fully_connected(net, 8)
net = tflearn.fully_connected(net, 8)
net = tflearn.fully_connected(net, len(output[0]), activation=softmax)
net = tflearn.regression(net)