This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| history = model.fit(X_train, y_train, | |
| batch_size=2048, epochs=150, | |
| callbacks=callbacks, | |
| validation_data=(X_valid, y_valid)) |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| from keras.callbacks import EarlyStopping, ModelCheckpoint | |
| # Create callbacks | |
| callbacks = [EarlyStopping(monitor='val_loss', patience=5), | |
| ModelCheckpoint('../models/model.h5'), save_best_only=True, | |
| save_weights_only=False)] |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| # Load in embeddings | |
| glove_vectors = '/home/ubuntu/.keras/datasets/glove.6B.100d.txt' | |
| glove = np.loadtxt(glove_vectors, dtype='str', comments=None) | |
| # Extract the vectors and words | |
| vectors = glove[:, 1:].astype('float') | |
| words = glove[:, 0] | |
| # Create lookup of words to vectors | |
| word_lookup = {word: vector for word, vector in zip(words, vectors)} |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| from keras.models import Sequential | |
| from keras.layers import LSTM, Dense, Dropout, Masking, Embedding | |
| model = Sequential() | |
| # Embedding layer | |
| model.add( | |
| Embedding(input_dim=num_words, | |
| input_length = training_length, | |
| output_dim=100, |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| features = [] | |
| labels = [] | |
| training_length = 50 | |
| # Iterate through the sequences of tokens | |
| for seq in sequences: | |
| # Create multiple training examples from each sequence | |
| for i in range(training_length, len(seq)): |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| from tpot import TPOTClassifier | |
| from sklearn.model_selection import TimeSeriesSplit | |
| # Cross validation object | |
| tss = TimeSeriesSplit(n_splits = 3) | |
| # Make the tpot search model | |
| tpot_pipeline = TPOTClassifier(generations = 10, population_size = 10, | |
| cv = tss, scoring = 'f1') | |
| # Find best model |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| from sklearn.ensemble import RandomForestClassifier | |
| # Create the model with 100 trees and default hyperparameters | |
| model = RandomForestClassifier(n_estimators=100) | |
| # Fit on training data | |
| model.fit(train, train_labels) | |
| # Make predictions on holdout test data | |
| predictions = model.predict(test) |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| import featuretools as ft | |
| # Primitives for deep feature synthesis | |
| trans_primitives = ['weekend', 'cum_sum', 'day', 'month', 'diff', 'time_since_previous'] | |
| agg_primitives = ['sum', 'time_since_last', 'avg_time_between', 'all', 'mode', | |
| 'num_unique', 'min', 'last', 'mean', 'percent_true', | |
| 'max', 'std', 'count'] | |
| # Perform deep feature synthesis | |
| feature_matrix, feature_names = ft.dfs(entityset=es, |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| import numpy as np | |
| import random | |
| random.seed(100) | |
| def generate_batch(pairs, n_positive = 50, negative_ratio = 1.0): | |
| """Generate batches of samples for training. | |
| Random select positive samples | |
| from pairs and randomly select negatives.""" | |
| # Create empty array to hold batch |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| __________________________________________________________________________________________________ | |
| Layer (type) Output Shape Param # Connected to | |
| ================================================================================================== | |
| book (InputLayer) (None, 1) 0 | |
| __________________________________________________________________________________________________ | |
| link (InputLayer) (None, 1) 0 | |
| __________________________________________________________________________________________________ | |
| book_embedding (Embedding) (None, 1, 50) 1851000 book[0][0] | |
| __________________________________________________________________________________________________ |