What is
self.label = tf.placeholder(
dtype=tf.float32,
shape=[self.batch_size, self.c_max_len, self.c_max_len],
name="label"
)
I'm a bit confused as to why the dimensions are context_length by context_length. A bit of context - I don't understand what you're doing in the following lines:
s_embedded = sentenceLSTM(sentences, real_lens, reuse = reuse)
c_embedded = tf.concat([s_embedded, labels], axis=1)
c_embedded = tf.reshape(c_embedded, shape = [self.batch_size, self.c_max_len, self.c_max_len + self.c_word_embed])
tagged_c_objects = tf.unstack(c_embedded, axis=1)
Could you explain this to me?
Cheers