如何使用经过 Keras 训练的嵌入式层?
How to use a Keras trained Embedded layer?
我的模型是:
model = Sequential()
model.add(Embedding(input_dim=vocab_size,
output_dim=1024, input_length=self.SEQ_LENGTH))
model.add(LSTM(vocab_size))
model.add(Dropout(rate=0.5))
model.add(Dense(vocab_size - 1, activation='softmax'))
而且我已经训练过了。但是现在在推理期间,我该如何使用该嵌入?
您的问题已解决here。作为骨架,您可以使用此代码:
from tensorflow.python.keras.preprocessing.text import Tokenizer
tokenizer_obj = Tokenizer()
tokenizer_obj.fit_on_texts(your_dataset)
...
max_length = max_number_words
X_test_tokens = tokenizer_obj.texts_to_sequences(X_test)
X_test_pad = pad_sequences(X_test_tokens, maxlen=max_length, padding='post')
score, acc = model.evaluate(X_test_pad, y_test, batch_size=128)
我的模型是:
model = Sequential()
model.add(Embedding(input_dim=vocab_size,
output_dim=1024, input_length=self.SEQ_LENGTH))
model.add(LSTM(vocab_size))
model.add(Dropout(rate=0.5))
model.add(Dense(vocab_size - 1, activation='softmax'))
而且我已经训练过了。但是现在在推理期间,我该如何使用该嵌入?
您的问题已解决here。作为骨架,您可以使用此代码:
from tensorflow.python.keras.preprocessing.text import Tokenizer
tokenizer_obj = Tokenizer()
tokenizer_obj.fit_on_texts(your_dataset)
...
max_length = max_number_words
X_test_tokens = tokenizer_obj.texts_to_sequences(X_test)
X_test_pad = pad_sequences(X_test_tokens, maxlen=max_length, padding='post')
score, acc = model.evaluate(X_test_pad, y_test, batch_size=128)