PKU_NLPDL_Assignment1 / task3 /embedding.py
Antoinegg1's picture
Upload folder using huggingface_hub
0ee6a96 verified
import gensim
from gensim.models import Word2Vec
import MeCab
def read_training_data(filepath):
english_sentences = []
japanese_sentences = []
with open(filepath, "r", encoding="utf-8") as f:
for line in f:
jp_sentence, en_sentence = line.strip().split("\t")
english_sentences.append(en_sentence.strip().split())
japanese_sentences.append(mecab.parse(jp_sentence).strip().split())
return english_sentences, japanese_sentences
def train_word2vec(sentences, language="en", embedding_size=300, window=5, min_count=1, sg=1):
# sg=1 skip-gram, sg=0 CBOW
model = Word2Vec(sentences, vector_size=embedding_size, window=window, min_count=min_count, sg=sg)
model.save(f"word2vec_{language}.model")
return model
if __name__ == "__main__":
mecab = MeCab.Tagger("-Owakati")
english_sentences, japanese_sentences = read_training_data("./data/train.txt")
en_model = train_word2vec(english_sentences, language="en", embedding_size=300, window=5, min_count=1, sg=0)
print(f"Trained English Word Embedding with {len(en_model.wv)} words.")
jp_model = train_word2vec(japanese_sentences, language="jp", embedding_size=300, window=5, min_count=1, sg=0)
print(f"Trained Japanese Word Embedding with {len(jp_model.wv)} words.")