ferdian15's picture
Create README.md
a63309f verified
from google.colab import drive
drive.mount('/content/drive')
import nltk
nltk.download('punkt')
nltk.download('wordnet')
import json
import random
import numpy as np
import nltk
from tensorflow.keras.models import Sequential
from tensorflow.keras.layers import Dense, Dropout
from tensorflow.keras.optimizers import SGD
from sklearn.preprocessing import LabelEncoder
from nltk.stem import WordNetLemmatizer
file_path = '/content/drive/MyDrive/Colab_Notebooks/Dataset/intents.json'
with open(file_path,'r') as file:
data = json.load(file)
lemmatizer = WordNetLemmatizer()
words = []
classes = []
documents = []
ignore_words = ['?', '!', '.']
for intent in data['intents']:
for pattern in intent['patterns']:
# Tokenize each word
word_list = nltk.word_tokenize(pattern)
words.extend(word_list)
documents.append((word_list, intent['tag']))
if intent['tag'] not in classes:
classes.append(intent['tag'])
words = [lemmatizer.lemmatize(w.lower()) for w in words if w not in ignore_words]
words = sorted(list(set(words)))
classes = sorted(list(set(classes)))
training = []
output_empty = [0] * len(classes)
for doc in documents:
bag = []
word_patterns = doc[0]
word_patterns = [lemmatizer.lemmatize(word.lower()) for word in word_patterns]
for w in words:
bag.append(1 if w in word_patterns else 0)
output_row = list(output_empty)
output_row[classes.index(doc[1])] = 1
training.append([bag, output_row])
random.shuffle(training)
training = np.array(training, dtype=object)
train_x = np.array(list(training[:, 0]))
train_y = np.array(list(training[:, 1]))
model = Sequential()
model.add(Dense(128, input_shape=(len(train_x[0]),), activation='relu'))
model.add(Dropout(0.5))
model.add(Dense(64, activation='relu'))
model.add(Dropout(0.5))
model.add(Dense(len(train_y[0]), activation='softmax'))
sgd = SGD(learning_rate=0.01, decay=1e-6, momentum=0.9, nesterov=True)
model.compile(loss='categorical_crossentropy', optimizer=sgd, metrics=['accuracy'])
hist = model.fit(train_x, train_y, epochs=200, batch_size=5, verbose=1)
model.save('chatbot_model.h5', hist)
print("Model created and saved successfully!")
import tensorflow as tf
model = tf.keras.models.load_model('chatbot_model.h5')
def clean_up_sentence(sentence):
sentence_words = nltk.word_tokenize(sentence)
sentence_words = [lemmatizer.lemmatize(word.lower()) for word in sentence_words]
return sentence_words
def bag_of_words(sentence, words):
sentence_words = clean_up_sentence(sentence)
bag = [0] * len(words)
for s in sentence_words:
for i, w in enumerate(words):
if w == s:
bag[i] = 1
return np.array(bag)
def predict_class(sentence, model):
bow = bag_of_words(sentence, words)
res = model.predict(np.array([bow]))[0]
ERROR_THRESHOLD = 0.25
results = [[i, r] for i, r in enumerate(res) if r > ERROR_THRESHOLD]
results.sort(key=lambda x: x[1], reverse=True)
return_list = [{"intent": classes[r[0]], "probability": str(r[1])} for r in results]
return return_list
def get_response(intents_list, intents_json):
tag = intents_list[0]['intent']
for i in intents_json['intents']:
if i['tag'] == tag:
return random.choice(i['responses'])
print("Bot is ready to chat! Type 'quit' to stop.")
while True:
message = input("You: ")
if message.lower() == "quit":
break
ints = predict_class(message, model)
if ints:
res = get_response(ints, data)
print("Bot:", res)
else:
print("Bot: Sorry, I don't understand that.")\