2017-07-12 10:25:55 +02:00
|
|
|
import json
|
|
|
|
import logging
|
2017-07-09 23:58:08 +02:00
|
|
|
import os
|
2017-07-03 13:48:12 +02:00
|
|
|
|
2017-07-12 10:25:55 +02:00
|
|
|
import numpy as np
|
2017-07-14 14:58:17 +02:00
|
|
|
import pandas as pd
|
|
|
|
import tensorflow as tf
|
2017-07-11 21:06:58 +02:00
|
|
|
from keras.callbacks import ModelCheckpoint, CSVLogger, EarlyStopping
|
2017-07-08 17:46:07 +02:00
|
|
|
from keras.models import load_model
|
2017-07-14 21:01:08 +02:00
|
|
|
from sklearn.decomposition import PCA
|
2017-07-14 15:57:52 +02:00
|
|
|
from sklearn.utils import class_weight
|
2017-06-30 10:12:20 +02:00
|
|
|
|
2017-07-12 10:25:55 +02:00
|
|
|
import arguments
|
2017-06-30 10:12:20 +02:00
|
|
|
import dataset
|
2017-07-07 16:48:10 +02:00
|
|
|
import hyperband
|
2017-06-30 10:12:20 +02:00
|
|
|
import models
|
2017-07-12 10:25:55 +02:00
|
|
|
# create logger
|
2017-07-14 14:58:17 +02:00
|
|
|
import visualize
|
|
|
|
from dataset import load_or_generate_h5data
|
|
|
|
|
2017-07-12 10:25:55 +02:00
|
|
|
logger = logging.getLogger('logger')
|
|
|
|
logger.setLevel(logging.DEBUG)
|
2017-07-03 13:48:12 +02:00
|
|
|
|
2017-07-12 10:25:55 +02:00
|
|
|
# create console handler and set level to debug
|
|
|
|
ch = logging.StreamHandler()
|
|
|
|
ch.setLevel(logging.DEBUG)
|
2017-07-05 21:19:19 +02:00
|
|
|
|
2017-07-12 10:25:55 +02:00
|
|
|
# create formatter
|
|
|
|
formatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')
|
2017-07-05 21:19:19 +02:00
|
|
|
|
2017-07-12 10:25:55 +02:00
|
|
|
# add formatter to ch
|
|
|
|
ch.setFormatter(formatter)
|
2017-07-03 13:48:12 +02:00
|
|
|
|
2017-07-12 10:25:55 +02:00
|
|
|
# add ch to logger
|
|
|
|
logger.addHandler(ch)
|
2017-07-05 21:19:19 +02:00
|
|
|
|
2017-07-12 10:25:55 +02:00
|
|
|
ch = logging.FileHandler("info.log")
|
|
|
|
ch.setLevel(logging.DEBUG)
|
2017-07-07 16:48:10 +02:00
|
|
|
|
2017-07-12 10:25:55 +02:00
|
|
|
# create formatter
|
|
|
|
formatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')
|
2017-07-03 13:48:12 +02:00
|
|
|
|
2017-07-12 10:25:55 +02:00
|
|
|
# add formatter to ch
|
|
|
|
ch.setFormatter(formatter)
|
2017-07-03 13:48:12 +02:00
|
|
|
|
2017-07-12 10:25:55 +02:00
|
|
|
# add ch to logger
|
|
|
|
logger.addHandler(ch)
|
2017-07-05 17:37:08 +02:00
|
|
|
|
2017-07-14 14:58:17 +02:00
|
|
|
print = logger.info
|
2017-07-08 15:04:58 +02:00
|
|
|
|
2017-07-14 14:58:17 +02:00
|
|
|
args = arguments.parse()
|
2017-07-05 17:37:08 +02:00
|
|
|
|
2017-07-14 14:58:17 +02:00
|
|
|
if args.gpu:
|
|
|
|
config = tf.ConfigProto(log_device_placement=True)
|
|
|
|
config.gpu_options.per_process_gpu_memory_fraction = 0.5
|
|
|
|
config.gpu_options.allow_growth = True
|
|
|
|
session = tf.Session(config=config)
|
2017-06-30 10:42:21 +02:00
|
|
|
|
|
|
|
|
2017-07-09 23:58:08 +02:00
|
|
|
def exists_or_make_path(p):
|
|
|
|
if not os.path.exists(p):
|
|
|
|
os.makedirs(p)
|
|
|
|
|
|
|
|
|
2017-07-08 11:53:03 +02:00
|
|
|
def main_paul_best():
|
|
|
|
char_dict = dataset.get_character_dict()
|
2017-07-14 14:58:17 +02:00
|
|
|
pauls_best_params = models.pauls_networks.best_config
|
|
|
|
pauls_best_params["vocab_size"] = len(char_dict) + 1
|
|
|
|
main_train(pauls_best_params)
|
2017-07-08 11:53:03 +02:00
|
|
|
|
|
|
|
|
2017-07-07 16:48:10 +02:00
|
|
|
def main_hyperband():
|
|
|
|
char_dict = dataset.get_character_dict()
|
|
|
|
|
|
|
|
params = {
|
|
|
|
# static params
|
|
|
|
"type": ["paul"],
|
2017-07-12 10:25:55 +02:00
|
|
|
"batch_size": [args.batch_size],
|
2017-07-07 16:48:10 +02:00
|
|
|
"vocab_size": [len(char_dict) + 1],
|
|
|
|
"window_size": [10],
|
|
|
|
"domain_length": [40],
|
|
|
|
"flow_features": [3],
|
|
|
|
"input_length": [40],
|
|
|
|
# model params
|
|
|
|
"embedding_size": [16, 32, 64, 128, 256, 512],
|
|
|
|
"filter_embedding": [16, 32, 64, 128, 256, 512],
|
|
|
|
"kernel_embedding": [1, 3, 5, 7, 9],
|
|
|
|
"hidden_embedding": [16, 32, 64, 128, 256, 512],
|
|
|
|
"dropout": [0.5],
|
|
|
|
"domain_features": [16, 32, 64, 128, 256, 512],
|
|
|
|
"filter_main": [16, 32, 64, 128, 256, 512],
|
|
|
|
"kernels_main": [1, 3, 5, 7, 9],
|
|
|
|
"dense_main": [16, 32, 64, 128, 256, 512],
|
|
|
|
}
|
|
|
|
|
2017-07-12 10:25:55 +02:00
|
|
|
logger.info("create training dataset")
|
2017-07-14 14:58:17 +02:00
|
|
|
domain_tr, flow_tr, client_tr, server_tr = load_or_generate_h5data(args.train_h5data, args.train_data,
|
2017-07-12 15:18:45 +02:00
|
|
|
args.domain_length, args.window)
|
2017-07-12 10:25:55 +02:00
|
|
|
hp = hyperband.Hyperband(params,
|
|
|
|
[domain_tr, flow_tr],
|
|
|
|
[client_tr, server_tr])
|
|
|
|
results = hp.run()
|
|
|
|
json.dump(results, open("hyperband.json"))
|
2017-07-07 16:48:10 +02:00
|
|
|
|
|
|
|
|
2017-07-14 15:57:52 +02:00
|
|
|
def get_custom_class_weights(client_tr, server_tr):
|
|
|
|
client = client_tr.value.argmax(1)
|
|
|
|
server = server_tr.value.argmax(1)
|
|
|
|
client_class_weight = class_weight.compute_class_weight('balanced', np.unique(client), client)
|
|
|
|
server_class_weight = class_weight.compute_class_weight('balanced', np.unique(server), server)
|
|
|
|
return {
|
|
|
|
"client": client_class_weight,
|
|
|
|
"server": server_class_weight
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2017-07-14 14:58:17 +02:00
|
|
|
def main_train(param=None):
|
2017-07-11 21:06:58 +02:00
|
|
|
exists_or_make_path(args.model_path)
|
|
|
|
|
|
|
|
char_dict = dataset.get_character_dict()
|
2017-07-14 14:58:17 +02:00
|
|
|
domain_tr, flow_tr, client_tr, server_tr = load_or_generate_h5data(args.train_h5data, args.train_data,
|
2017-07-11 21:06:58 +02:00
|
|
|
args.domain_length, args.window)
|
2017-06-30 10:12:20 +02:00
|
|
|
|
2017-07-09 23:58:08 +02:00
|
|
|
# parameter
|
2017-07-14 14:58:17 +02:00
|
|
|
p = {
|
2017-07-14 21:01:08 +02:00
|
|
|
"type": args.model_type,
|
2017-07-09 23:58:08 +02:00
|
|
|
"batch_size": 64,
|
|
|
|
"window_size": args.window,
|
|
|
|
"domain_length": args.domain_length,
|
|
|
|
"flow_features": 3,
|
|
|
|
"vocab_size": len(char_dict) + 1,
|
|
|
|
#
|
|
|
|
'dropout': 0.5,
|
|
|
|
'domain_features': args.domain_embedding,
|
|
|
|
'embedding_size': args.embedding,
|
|
|
|
'filter_main': 128,
|
|
|
|
'flow_features': 3,
|
2017-07-14 21:01:08 +02:00
|
|
|
# 'dense_main': 512,
|
|
|
|
'dense_main': 128,
|
2017-07-09 23:58:08 +02:00
|
|
|
'filter_embedding': args.hidden_char_dims,
|
|
|
|
'hidden_embedding': args.domain_embedding,
|
|
|
|
'kernel_embedding': 3,
|
|
|
|
'kernels_main': 3,
|
|
|
|
'input_length': 40
|
|
|
|
}
|
2017-07-14 14:58:17 +02:00
|
|
|
if not param:
|
|
|
|
param = p
|
2017-06-30 10:12:20 +02:00
|
|
|
|
2017-07-09 23:58:08 +02:00
|
|
|
embedding, model = models.get_models_by_params(param)
|
2017-07-08 11:53:03 +02:00
|
|
|
embedding.summary()
|
2017-07-05 17:37:08 +02:00
|
|
|
model.summary()
|
2017-07-12 10:25:55 +02:00
|
|
|
logger.info("define callbacks")
|
2017-07-14 14:58:17 +02:00
|
|
|
callbacks = []
|
|
|
|
callbacks.append(ModelCheckpoint(filepath=args.clf_model,
|
|
|
|
monitor='val_loss',
|
|
|
|
verbose=False,
|
|
|
|
save_best_only=True))
|
|
|
|
callbacks.append(CSVLogger(args.train_log))
|
|
|
|
if args.stop_early:
|
|
|
|
callbacks.append(EarlyStopping(monitor='val_loss',
|
|
|
|
patience=5,
|
|
|
|
verbose=False))
|
2017-07-12 10:25:55 +02:00
|
|
|
logger.info("compile model")
|
2017-07-14 14:58:17 +02:00
|
|
|
custom_metrics = models.get_metric_functions()
|
2017-06-30 10:12:20 +02:00
|
|
|
model.compile(optimizer='adam',
|
2017-07-07 16:48:10 +02:00
|
|
|
loss='categorical_crossentropy',
|
2017-07-14 14:58:17 +02:00
|
|
|
metrics=['accuracy'] + custom_metrics)
|
2017-07-14 15:57:52 +02:00
|
|
|
|
|
|
|
if args.class_weights:
|
|
|
|
logger.info("class weights: compute custom weights")
|
|
|
|
custom_class_weights = get_custom_class_weights(client_tr, server_tr)
|
2017-07-14 21:01:08 +02:00
|
|
|
logger.info(custom_class_weights)
|
2017-07-14 15:57:52 +02:00
|
|
|
else:
|
|
|
|
logger.info("class weights: set default")
|
|
|
|
custom_class_weights = None
|
2017-07-12 10:25:55 +02:00
|
|
|
logger.info("start training")
|
2017-07-05 18:37:29 +02:00
|
|
|
model.fit([domain_tr, flow_tr],
|
2017-07-05 21:19:19 +02:00
|
|
|
[client_tr, server_tr],
|
2017-07-03 13:48:12 +02:00
|
|
|
batch_size=args.batch_size,
|
|
|
|
epochs=args.epochs,
|
2017-07-14 14:58:17 +02:00
|
|
|
callbacks=callbacks,
|
2017-07-05 21:19:19 +02:00
|
|
|
shuffle=True,
|
2017-07-14 15:57:52 +02:00
|
|
|
validation_split=0.2,
|
|
|
|
class_weight=custom_class_weights)
|
2017-07-12 10:25:55 +02:00
|
|
|
logger.info("save embedding")
|
2017-07-08 15:04:58 +02:00
|
|
|
embedding.save(args.embedding_model)
|
2017-07-07 16:48:10 +02:00
|
|
|
|
2017-07-05 21:19:19 +02:00
|
|
|
|
2017-07-06 16:27:47 +02:00
|
|
|
def main_test():
|
2017-07-14 14:58:17 +02:00
|
|
|
domain_val, flow_val, client_val, server_val = load_or_generate_h5data(args.test_h5data, args.test_data,
|
2017-07-11 21:06:58 +02:00
|
|
|
args.domain_length, args.window)
|
2017-07-14 14:58:17 +02:00
|
|
|
clf = load_model(args.clf_model, custom_objects=models.get_metrics())
|
2017-07-14 15:57:52 +02:00
|
|
|
# stats = clf.evaluate([domain_val, flow_val],
|
|
|
|
# [client_val, server_val],
|
|
|
|
# batch_size=args.batch_size)
|
2017-07-12 10:25:55 +02:00
|
|
|
y_pred = clf.predict([domain_val, flow_val],
|
2017-07-16 09:42:52 +02:00
|
|
|
batch_size=args.batch_size,
|
|
|
|
verbose=1)
|
2017-07-14 14:58:17 +02:00
|
|
|
np.save(args.future_prediction, y_pred)
|
2017-06-30 10:12:20 +02:00
|
|
|
|
2017-07-17 19:30:56 +02:00
|
|
|
char_dict = dataset.get_character_dict()
|
|
|
|
user_flow_df = dataset.get_user_flow_data(args.test_data)
|
|
|
|
domains = user_flow_df.domain.unique()
|
|
|
|
|
|
|
|
def get_domain_features_reduced(d):
|
|
|
|
return dataset.get_domain_features(d[0], char_dict, args.domain_length)
|
|
|
|
|
|
|
|
domain_features = []
|
|
|
|
for ds in domains:
|
|
|
|
domain_features.append(np.apply_along_axis(get_domain_features_reduced, 2, np.atleast_3d(ds)))
|
|
|
|
|
|
|
|
model = load_model(args.embedding_model)
|
|
|
|
domain_features = np.stack(domain_features).reshape((-1, 40))
|
|
|
|
pred = model.predict(domains, batch_size=args.batch_size, verbose=1)
|
|
|
|
|
|
|
|
np.save("/tmp/rk/domains.npy", domains)
|
|
|
|
np.save("/tmp/rk/domain_features.npy", domain_features)
|
|
|
|
np.save("/tmp/rk/domain_embd.npy", pred)
|
|
|
|
|
2017-06-30 10:12:20 +02:00
|
|
|
|
2017-07-07 08:43:16 +02:00
|
|
|
def main_visualization():
|
2017-07-14 21:01:08 +02:00
|
|
|
domain_val, flow_val, client_val, server_val = load_or_generate_h5data(args.test_h5data, args.test_data,
|
|
|
|
args.domain_length, args.window)
|
2017-07-12 10:25:55 +02:00
|
|
|
logger.info("plot model")
|
2017-07-14 14:58:17 +02:00
|
|
|
model = load_model(args.clf_model, custom_objects=models.get_metrics())
|
2017-07-14 21:01:08 +02:00
|
|
|
visualize.plot_model(model, os.path.join(args.model_path, "model.png"))
|
2017-07-17 19:30:56 +02:00
|
|
|
try:
|
|
|
|
logger.info("plot training curve")
|
|
|
|
logs = pd.read_csv(args.train_log)
|
|
|
|
visualize.plot_training_curve(logs, "client", "{}/client_train.png".format(args.model_path))
|
|
|
|
visualize.plot_training_curve(logs, "server", "{}/server_train.png".format(args.model_path))
|
|
|
|
except Exception as e:
|
|
|
|
logger.warning(f"could not generate training curves: {e}")
|
2017-07-14 14:58:17 +02:00
|
|
|
|
|
|
|
client_pred, server_pred = np.load(args.future_prediction)
|
2017-07-12 10:25:55 +02:00
|
|
|
logger.info("plot pr curve")
|
2017-07-14 14:58:17 +02:00
|
|
|
visualize.plot_precision_recall(client_val.value, client_pred, "{}/client_prc.png".format(args.model_path))
|
|
|
|
visualize.plot_precision_recall(server_val.value, server_pred, "{}/server_prc.png".format(args.model_path))
|
|
|
|
visualize.plot_precision_recall_curves(client_val.value, client_pred, "{}/client_prc2.png".format(args.model_path))
|
|
|
|
visualize.plot_precision_recall_curves(server_val.value, server_pred, "{}/server_prc2.png".format(args.model_path))
|
2017-07-12 10:25:55 +02:00
|
|
|
logger.info("plot roc curve")
|
2017-07-14 14:58:17 +02:00
|
|
|
visualize.plot_roc_curve(client_val.value, client_pred, "{}/client_roc.png".format(args.model_path))
|
|
|
|
visualize.plot_roc_curve(server_val.value, server_pred, "{}/server_roc.png".format(args.model_path))
|
2017-07-14 15:57:52 +02:00
|
|
|
visualize.plot_confusion_matrix(client_val.value.argmax(1), client_pred.argmax(1),
|
|
|
|
"{}/client_cov.png".format(args.model_path),
|
|
|
|
normalize=False, title="Client Confusion Matrix")
|
|
|
|
visualize.plot_confusion_matrix(server_val.value.argmax(1), server_pred.argmax(1),
|
|
|
|
"{}/server_cov.png".format(args.model_path),
|
|
|
|
normalize=False, title="Server Confusion Matrix")
|
2017-07-07 08:43:16 +02:00
|
|
|
|
2017-07-14 21:01:08 +02:00
|
|
|
# embedding visi
|
|
|
|
import matplotlib.pyplot as plt
|
|
|
|
|
|
|
|
model = load_model(args.embedding_model)
|
2017-07-17 19:30:56 +02:00
|
|
|
domains = np.reshape(domain_val, (domain_val.shape[0] * domain_val.shape[1], 40))
|
|
|
|
domain_embedding = model.predict(domains, batch_size=args.batch_size, verbose=1)
|
2017-07-14 21:01:08 +02:00
|
|
|
|
|
|
|
pca = PCA(n_components=2)
|
|
|
|
domain_reduced = pca.fit_transform(domain_embedding)
|
|
|
|
print(pca.explained_variance_ratio_)
|
|
|
|
|
|
|
|
clients = np.repeat(client_val, 10, axis=0)
|
|
|
|
clients = clients.argmax(1)
|
|
|
|
servers = np.repeat(server_val, 10, axis=0)
|
|
|
|
servers = servers.argmax(1)
|
|
|
|
|
|
|
|
plt.scatter(domain_reduced[:, 0], domain_reduced[:, 1], c=clients, cmap=plt.cm.bwr, s=2)
|
|
|
|
plt.show()
|
|
|
|
plt.scatter(domain_reduced[:, 0], domain_reduced[:, 1], c=servers, cmap=plt.cm.bwr, s=2)
|
|
|
|
plt.show()
|
2017-07-07 08:43:16 +02:00
|
|
|
|
2017-07-16 09:42:52 +02:00
|
|
|
|
2017-07-07 08:43:16 +02:00
|
|
|
def main_score():
|
2017-07-14 14:58:17 +02:00
|
|
|
# mask = dataset.load_mask_eval(args.data, args.test_image)
|
|
|
|
# pred = np.load(args.pred)
|
|
|
|
# visualize.score_model(mask, pred)
|
|
|
|
pass
|
2017-07-07 08:43:16 +02:00
|
|
|
|
|
|
|
|
2017-07-16 09:42:52 +02:00
|
|
|
def main_data():
|
|
|
|
char_dict = dataset.get_character_dict()
|
|
|
|
user_flow_df = dataset.get_user_flow_data(args.train_data)
|
|
|
|
logger.info("create training dataset")
|
|
|
|
domain_tr, flow_tr, client_tr, server_tr = dataset.create_dataset_from_flows(user_flow_df, char_dict,
|
|
|
|
max_len=args.domain_length,
|
|
|
|
window_size=args.window)
|
|
|
|
print(f"domain shape {domain_tr.shape}")
|
|
|
|
print(f"flow shape {flow_tr.shape}")
|
|
|
|
print(f"client shape {client_tr.shape}")
|
|
|
|
print(f"server shape {server_tr.shape}")
|
|
|
|
|
|
|
|
|
2017-07-07 08:43:16 +02:00
|
|
|
def main():
|
2017-07-07 16:48:10 +02:00
|
|
|
if "train" in args.modes:
|
|
|
|
main_train()
|
|
|
|
if "hyperband" in args.modes:
|
|
|
|
main_hyperband()
|
|
|
|
if "test" in args.modes:
|
|
|
|
main_test()
|
|
|
|
if "fancy" in args.modes:
|
|
|
|
main_visualization()
|
|
|
|
if "score" in args.modes:
|
|
|
|
main_score()
|
2017-07-08 11:53:03 +02:00
|
|
|
if "paul" in args.modes:
|
|
|
|
main_paul_best()
|
2017-07-16 09:42:52 +02:00
|
|
|
if "data" in args.modes:
|
|
|
|
main_data()
|
2017-07-07 08:43:16 +02:00
|
|
|
|
|
|
|
|
2017-06-30 10:12:20 +02:00
|
|
|
if __name__ == "__main__":
|
|
|
|
main()
|