System information
What is the top-level directory of the model you are using: object_detection/ssd_inception_v2 together with slim/InceptionV2
Have I written custom code (as opposed to using a stock example script provided in TensorFlow): No
I trained an inceptionV2 (from slim) on a (traffic sign) classification dataset. I then wanted to use the resulting checkpoint as a basis for the object detection API. If I understand correctly, it should be possible to just use and adapted version of the /object_detection/samples/configs/ ssd_inception_v2_pets.config config file. However when I run the object detection
python object_detection/train.py --logtostderr --pipeline_config_path=/home/tobi/tensorflow/trafficsigns/models/model/ssd_inception_v2_trafficsigns.config –train_dir=/home/tobi/tensorflow/trafficsigns/models/model/train
I get the following error
Traceback (most recent call last):
File "object_detection/train.py", line 198, in <module>
tf.app.run()
File "/home/tobi/tensorflow/local/lib/python2.7/site-packages/tensorflow/python/platform/app.py", line 48, in run
_sys.exit(main(_sys.argv[:1] + flags_passthrough))
File "object_detection/train.py", line 194, in main
worker_job_name, is_chief, FLAGS.train_dir)
File "/home/tobi/tensorflow/lib/python2.7/site-packages/tensorflow/models/object_detection/trainer.py", line 191, in train
clones = model_deploy.create_clones(deploy_config, model_fn, [input_queue])
File "/home/tobi/tensorflow/local/lib/python2.7/site-packages/tensorflow/models/slim/deployment/model_deploy.py", line 193, in create_clones
outputs = model_fn(*args, **kwargs)
File "/home/tobi/tensorflow/lib/python2.7/site-packages/tensorflow/models/object_detection/trainer.py", line 132, in _create_losses
losses_dict = detection_model.loss(prediction_dict)
File "/home/tobi/tensorflow/lib/python2.7/site-packages/tensorflow/models/object_detection/meta_architectures/ssd_meta_arch.py", line 431, in loss
location_losses, cls_losses, prediction_dict, match_list)
File "/home/tobi/tensorflow/lib/python2.7/site-packages/tensorflow/models/object_detection/meta_architectures/ssd_meta_arch.py", line 551, in _apply_hard_mining
[0, 0, 1], class_pred_shape), class_pred_shape)
File "/home/tobi/tensorflow/local/lib/python2.7/site-packages/tensorflow/python/ops/array_ops.py", line 547, in slice
return gen_array_ops._slice(input_, begin, size, name=name)
File "/home/tobi/tensorflow/local/lib/python2.7/site-packages/tensorflow/python/ops/gen_array_ops.py", line 2896, in _slice
name=name)
File "/home/tobi/tensorflow/local/lib/python2.7/site-packages/tensorflow/python/framework/op_def_library.py", line 499, in apply_op
repr(values), type(values).__name__))
TypeError: Expected int32 passed to parameter 'size' of op 'Slice', got [-1, None, 43] of type 'list' instead.
It seems to me the problem is that the anchor boxes are not initialized resulting in the ‘None’ value which lets the tf.Slice operation crash.
The config file I used
# SSD with Inception v2 configured for Oxford-IIIT Pets Dataset.
# Users should configure the fine_tune_checkpoint field in the train config as
# well as the label_map_path and input_path fields in the train_input_reader and
# eval_input_reader. Search for "PATH_TO_BE_CONFIGURED" to find the fields that
# should be configured.
model {
ssd {
num_classes: 43
box_coder {
faster_rcnn_box_coder {
y_scale: 10.0
x_scale: 10.0
height_scale: 5.0
width_scale: 5.0
}
}
matcher {
argmax_matcher {
matched_threshold: 0.5
unmatched_threshold: 0.5
ignore_thresholds: false
negatives_lower_than_unmatched: true
force_match_for_each_row: true
}
}
similarity_calculator {
iou_similarity {
}
}
anchor_generator {
ssd_anchor_generator {
num_layers: 6
min_scale: 0.2
max_scale: 0.95
aspect_ratios: 1.0
aspect_ratios: 2.0
aspect_ratios: 0.5
aspect_ratios: 3.0
aspect_ratios: 0.3333
reduce_boxes_in_lowest_layer: true
}
}
image_resizer {
keep_aspect_ratio_resizer {
min_dimension: 400
max_dimension: 690
}
}
box_predictor {
convolutional_box_predictor {
min_depth: 0
max_depth: 0
num_layers_before_predictor: 0
use_dropout: false
dropout_keep_probability: 0.8
kernel_size: 3
box_code_size: 4
apply_sigmoid_to_scores: false
conv_hyperparams {
activation: RELU_6,
regularizer {
l2_regularizer {
weight: 0.00004
}
}
initializer {
truncated_normal_initializer {
stddev: 0.03
mean: 0.0
}
}
}
}
}
feature_extractor {
type: 'ssd_inception_v2'
min_depth: 16
depth_multiplier: 1.0
conv_hyperparams {
activation: RELU_6,
regularizer {
l2_regularizer {
weight: 0.00004
}
}
initializer {
truncated_normal_initializer {
stddev: 0.03
mean: 0.0
}
}
batch_norm {
train: true,
scale: true,
center: true,
decay: 0.9997,
epsilon: 0.001,
}
}
}
loss {
classification_loss {
weighted_sigmoid {
anchorwise_output: true
}
}
localization_loss {
weighted_smooth_l1 {
anchorwise_output: true
}
}
hard_example_miner {
num_hard_examples: 3000
iou_threshold: 0.99
loss_type: CLASSIFICATION
max_negatives_per_positive: 3
min_negatives_per_image: 0
}
classification_weight: 1.0
localization_weight: 1.0
}
normalize_loss_by_num_matches: true
post_processing {
batch_non_max_suppression {
score_threshold: 1e-8
iou_threshold: 0.6
max_detections_per_class: 100
max_total_detections: 100
}
score_converter: SIGMOID
}
}
}
train_config: {
batch_size: 24
optimizer {
rms_prop_optimizer: {
learning_rate: {
exponential_decay_learning_rate {
initial_learning_rate: 0.004
decay_steps: 800720
decay_factor: 0.95
}
}
momentum_optimizer_value: 0.9
decay: 0.9
epsilon: 1.0
}
}
fine_tune_checkpoint: "/home/tobi/tensorflow/trafficsigns/models/pretrained/inception_v2_GTS/model.ckpt"
from_detection_checkpoint: false
# Note: The below line limits the training process to 200K steps, which we
# empirically found to be sufficient enough to train the pets dataset. This
# effectively bypasses the learning rate schedule (the learning rate will
# never decay). Remove the below line to train indefinitely.
#num_steps: 200000
data_augmentation_options {
random_horizontal_flip {
}
}
data_augmentation_options {
ssd_random_crop {
}
}
}
train_input_reader: {
tf_record_input_reader {
input_path: "/home/tobi/tensorflow/trafficsigns/data/train.record"
}
label_map_path: "/home/tobi/tensorflow/trafficsigns/data/ts_label_map.pbtxt"
}
eval_config: {
num_examples: 2000
num_visualizations: 25
# Note: The below line limits the evaluation process to 10 evaluations.
# Remove the below line to evaluate indefinitely.
max_evals: 10
}
eval_input_reader: {
tf_record_input_reader {
input_path: "/home/tobi/tensorflow/trafficsigns/data/test.record"
}
label_map_path: "/home/tobi/tensorflow/trafficsigns/data/ts_label_map.pbtxt"
shuffle: false
num_readers: 1
}
Any ideas what went wrong? Thank you in advance.
Edit
After changing the keep_aspect_ratio_resizer to the fixed_shape_resizer in the config file as suggested (thx again) the training process started. I somehow get twice the logging messages which does not further bother me. What concerns me are the following warnings that indicate that the checkpoint does not contain expected variables. The training works anyway, I have not checked the results yet but I am not sure if I should ignore these.
WARNING:root:Variable [InceptionV2/Conv2d_1a_7x7/BatchNorm/beta] not available in checkpoint
WARNING:root:Variable [InceptionV2/Conv2d_1a_7x7/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Conv2d_1a_7x7/BatchNorm/moving_mean] not available in checkpoint
WARNING:root:Variable [InceptionV2/Conv2d_1a_7x7/BatchNorm/moving_variance] not available in checkpoint
WARNING:root:Variable [InceptionV2/Conv2d_2b_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Conv2d_2c_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_3b/Branch_0/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_3b/Branch_1/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_3b/Branch_1/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_3b/Branch_2/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_3b/Branch_2/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_3b/Branch_2/Conv2d_0c_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_3b/Branch_3/Conv2d_0b_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_3c/Branch_0/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_3c/Branch_1/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_3c/Branch_1/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_3c/Branch_2/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_3c/Branch_2/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_3c/Branch_2/Conv2d_0c_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_3c/Branch_3/Conv2d_0b_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4a/Branch_0/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4a/Branch_0/Conv2d_1a_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4a/Branch_1/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4a/Branch_1/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4a/Branch_1/Conv2d_1a_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4b/Branch_0/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4b/Branch_1/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4b/Branch_1/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4b/Branch_2/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4b/Branch_2/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4b/Branch_2/Conv2d_0c_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4b/Branch_3/Conv2d_0b_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4c/Branch_0/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4c/Branch_1/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4c/Branch_1/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4c/Branch_2/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4c/Branch_2/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4c/Branch_2/Conv2d_0c_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4c/Branch_3/Conv2d_0b_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4d/Branch_0/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4d/Branch_1/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4d/Branch_1/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4d/Branch_2/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4d/Branch_2/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4d/Branch_2/Conv2d_0c_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4d/Branch_3/Conv2d_0b_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4e/Branch_0/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4e/Branch_1/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4e/Branch_1/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4e/Branch_2/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4e/Branch_2/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4e/Branch_2/Conv2d_0c_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_4e/Branch_3/Conv2d_0b_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5a/Branch_0/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5a/Branch_0/Conv2d_1a_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5a/Branch_1/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5a/Branch_1/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5a/Branch_1/Conv2d_1a_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5b/Branch_0/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5b/Branch_1/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5b/Branch_1/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5b/Branch_2/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5b/Branch_2/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5b/Branch_2/Conv2d_0c_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5b/Branch_3/Conv2d_0b_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c/Branch_0/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c/Branch_1/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c/Branch_1/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c/Branch_2/Conv2d_0a_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c/Branch_2/Conv2d_0b_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c/Branch_2/Conv2d_0c_3x3/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c/Branch_3/Conv2d_0b_1x1/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_2_1x1_256/BatchNorm/beta] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_2_1x1_256/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_2_1x1_256/BatchNorm/moving_mean] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_2_1x1_256/BatchNorm/moving_variance] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_2_1x1_256/weights] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_3_1x1_128/BatchNorm/beta] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_3_1x1_128/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_3_1x1_128/BatchNorm/moving_mean] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_3_1x1_128/BatchNorm/moving_variance] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_3_1x1_128/weights] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_4_1x1_128/BatchNorm/beta] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_4_1x1_128/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_4_1x1_128/BatchNorm/moving_mean] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_4_1x1_128/BatchNorm/moving_variance] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_4_1x1_128/weights] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_5_1x1_64/BatchNorm/beta] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_5_1x1_64/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_5_1x1_64/BatchNorm/moving_mean] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_5_1x1_64/BatchNorm/moving_variance] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_1_Conv2d_5_1x1_64/weights] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_2_3x3_s2_512/BatchNorm/beta] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_2_3x3_s2_512/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_2_3x3_s2_512/BatchNorm/moving_mean] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_2_3x3_s2_512/BatchNorm/moving_variance] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_2_3x3_s2_512/weights] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_3_3x3_s2_256/BatchNorm/beta] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_3_3x3_s2_256/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_3_3x3_s2_256/BatchNorm/moving_mean] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_3_3x3_s2_256/BatchNorm/moving_variance] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_3_3x3_s2_256/weights] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_4_3x3_s2_256/BatchNorm/beta] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_4_3x3_s2_256/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_4_3x3_s2_256/BatchNorm/moving_mean] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_4_3x3_s2_256/BatchNorm/moving_variance] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_4_3x3_s2_256/weights] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_5_3x3_s2_128/BatchNorm/beta] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_5_3x3_s2_128/BatchNorm/gamma] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_5_3x3_s2_128/BatchNorm/moving_mean] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_5_3x3_s2_128/BatchNorm/moving_variance] not available in checkpoint
WARNING:root:Variable [InceptionV2/Mixed_5c_2_Conv2d_5_3x3_s2_128/weights] not available in checkpoint
I see at least one problem with the above config --- which is that the keep_aspect_ratio_resizer is incompatible with batch sizes greater than 1. (I know this is not documented anywhere yet). The problem is that after resizing each image, because the aspect ratios are preserved, all the images have a different shape and we then aren't able to stack them into a batch.
I recommend using a fixed shape like 300x300.