xtr33me
xtr33me

Reputation: 1106

Why am I getting shape errors when trying to pass a batch from the Tensorflow Dataset API to my session operations?

I am dealing with an issue in my conversion over to the Dataset API and I guess I just don't have enough experience yet with the API to know how to handle the below situation. We currently have image augmentation that we perform currently using queueing and batching. I was tasked with checking out the new Dataset API and converting over our existing implementation using it rather than queues.

What we would like to do is get a reference to all the paths and handle all operations from just that reference. As you see in the dataset initialization, I have mapped the parse_fn to the dataset itself which then goes about reading the file and extracting the initial values from the filenames. However when I then go about calling the iterators next_batch method and then pass those values to get_summary, I'm now getting an error around shape. I have been trying a number of things which just keeps changing the error and so I felt I should see if anyone on SO saw possibly that I was going about this all wrong and should be taking a different route. Does anything jump out as absolutely wrong in my use of the Dataset API?

Should I not be calling the ops this way any longer? I noticed the majority of the examples I saw they would get the batch, pass the variables to the op and then capture that in a variable and pass that to sess.run, however I haven't found an easy way of doing that as of yet with our setup that wasn't erroring so this was the approach I took instead (but its still erroring). I'll be continuing to try to trace down the problem and post here should I find anything, but if anyone sees something please advise. Thanks!

Current Error:

... in get_summary summary, acc = sess.run([self._summary_op, self._accuracy], feed_dict=feed_dict) ValueError: Cannot feed value of shape (32,) for Tensor 'ph_input_labels:0', which has shape '(?, 1)

Below is the block where the get_summary method is called and error is fired:

def perform_train():
    if __name__ == '__main__':
        #Get all our image paths
        filenames = data_layer_train.get_image_paths()
        next_batch, iterator = preproc_image_fn(filenames=filenames)

    with tf.Session(config=tf.ConfigProto(gpu_options=gpu_options)) as sess:
        with sess.graph.as_default():
            # Set the random seed for tensorflow
            tf.set_random_seed(cfg.RNG_SEED)

            classifier_network = c_common.create_model(len(products_to_class_dict), is_training=True)
            optimizer, global_step_var = c_common.create_optimizer(classifier_network)

            sess.run(tf.local_variables_initializer())
            sess.run(tf.global_variables_initializer())

            # Init tables and dataset iterator
            sess.run(tf.tables_initializer())
            sess.run(iterator.initializer)

            cur_epoch = 0
            blobs = None
            try:
                epoch_size = data_layer_train.get_steps_per_epoch()
                num_steps = num_epochs * epoch_size
                for step in range(num_steps):
                    timer_summary.tic()
                    if blobs is None:
                        #Now populate from our training dataset
                        blobs = sess.run(next_batch)

                    # *************** Below is where it is erroring *****************
                    summary_train, acc = classifier_network.get_summary(sess, blobs["images"], blobs["labels"], blobs["weights"])

            ...

Believe the error is in preproc_image_fn:

def preproc_image_fn(filenames, images=None, labels=None, image_paths=None, cells=None, weights=None):
    def _parse_fn(filename, label, weight):
        augment_instance = False
        paths=[]
        selected_cells=[]
        if vals.FIRST_ITER:
            #Perform our check of the path to see if _data_augmentation is within it
            #If so set augment_instance to true and replace the substring with an empty string
            new_filename = tf.regex_replace(filename, "_data_augmentation", "")
            contains = tf.equal(tf.size(tf.string_split([filename], "")), tf.size(tf.string_split([new_filename])))
            filename = new_filename
            if contains is True:
                augment_instance = True

        core_file = tf.string_split([filename], '\\').values[-1]
        product_id = tf.string_split([core_file], ".").values[0]

        label = search_tf_table_for_entry(product_id)
        weight = data_layer_train.get_weights(product_id)

        image_string = tf.read_file(filename)
        img = tf.image.decode_image(image_string, channels=data_layer_train._channels)
        img.set_shape([None, None, None])
        img = tf.image.resize_images(img, [data_layer_train._target_height, data_layer_train._target_width])
        #Previously I was returning the below, but I was getting an error from the op when assigning feed_dict stating that it didnt like the dictionary
        #retval = dict(zip([filename], [img])), label, weight
        retval = img, label, weight
        return retval

    num_files = len(filenames)
    filenames = tf.constant(filenames)

    #*********** Setup dataset below ************
    dataset = tf.data.Dataset.from_tensor_slices((filenames, labels, weights))
    dataset=dataset.map(_parse_fn)
    dataset = dataset.repeat()
    dataset = dataset.batch(32)
    iterator = dataset.make_initializable_iterator()

    batch_features,  batch_labels, batch_weights = iterator.get_next()
    return {'images': batch_features, 'labels': batch_labels, 'weights': batch_weights}, iterator

def search_tf_table_for_entry(self, product_id):
    '''Looks up keys in the table and outputs the values. Will return -1 if not found '''
    if product_id is not None:
        return self._products_to_class_table.lookup(product_id)
    else:
        if not self._real_eval:
            logger().info("class not found in training {} ".format(product_id))
        return -1

Where I create the model and have the placeholders used previously:

...
 def create_model(self):
    weights_regularizer = tf.contrib.layers.l2_regularizer(cfg.TRAIN.WEIGHT_DECAY)
    biases_regularizer = weights_regularizer

    # Input data.
    self._input_images = tf.placeholder(
        tf.float32, shape=(None, self._image_height, self._image_width, self._num_channels), name="ph_input_images")
    self._input_labels = tf.placeholder(tf.int64, shape=(None, 1), name="ph_input_labels")
    self._input_weights = tf.placeholder(tf.float32, shape=(None, 1), name="ph_input_weights")
    self._is_training = tf.placeholder(tf.bool, name='ph_is_training')
    self._keep_prob = tf.placeholder(tf.float32, name="ph_keep_prob")
    self._accuracy = tf.reduce_mean(tf.cast(self._correct_prediction, tf.float32))
    ...
    self.create_summaries()

def create_summaries(self):
    val_summaries = []
    with tf.device("/cpu:0"):
        for var in self._act_summaries:
            self._add_act_summary(var)
        for var in self._train_summaries:
            self._add_train_summary(var)

    self._summary_op = tf.summary.merge_all()
    self._summary_op_val = tf.summary.merge(val_summaries)

def get_summary(self, sess, images, labels, weights):
    feed_dict = {self._input_images: images, self._input_labels: labels,
                 self._input_weights: weights, self._is_training: False}

    summary, acc = sess.run([self._summary_op, self._accuracy], feed_dict=feed_dict)

    return summary, acc

Upvotes: 0

Views: 182

Answers (1)

Y. Luo
Y. Luo

Reputation: 5732

Since the error says:

Cannot feed value of shape (32,) for Tensor 'ph_input_labels:0', which has shape '(?, 1)

My guess is your labels in get_summary has the shape [32]. Can you just reshape it to (32, 1)? Or maybe reshape the label earlier in _parse_fn?

Upvotes: 1

Related Questions