fix flat 2 model bug caused by changing input layer
This commit is contained in:
parent
5741f8ee0e
commit
0b26c6125c
@ -35,7 +35,7 @@ parser.add_argument("--type", action="store", dest="model_type",
|
||||
default="final") # inter, final, staggered
|
||||
|
||||
parser.add_argument("--depth", action="store", dest="model_depth",
|
||||
default="small") # small, medium
|
||||
default="flat1") # small, medium
|
||||
|
||||
parser.add_argument("--model_output", action="store", dest="model_output",
|
||||
default="both")
|
||||
|
@ -41,9 +41,9 @@ def get_model(cnnDropout, flow_features, domain_features, window_size, domain_le
|
||||
ipt_domains = Input(shape=(window_size, domain_length), name="ipt_domains")
|
||||
encoded = TimeDistributed(cnn, name="domain_cnn")(ipt_domains)
|
||||
ipt_flows = Input(shape=(window_size, flow_features), name="ipt_flows")
|
||||
ipt_flows = BatchNormalization()(ipt_flows)
|
||||
ipt_flows = Dense(dense_dim, activation=selu)(ipt_flows)
|
||||
merged = keras.layers.concatenate([encoded, ipt_flows], -1)
|
||||
y = BatchNormalization()(ipt_flows)
|
||||
y = Dense(dense_dim, activation=selu)(y)
|
||||
merged = keras.layers.concatenate([encoded, y], -1)
|
||||
# CNN processing a small slides of flow windows
|
||||
y = Conv1D(cnn_dims,
|
||||
kernel_size,
|
||||
|
Loading…
Reference in New Issue
Block a user