kolovoza 26, 2020 —
Posted by Kensen Shi, Google Research
When manipulating tensors, one must keep track of multiple dimensions, tensor shape and DType compatibility, and of course mathematical correctness. Additionally, there are hundreds of TensorFlow operations, and finding the right ones to use can be a challenge.
Instead of coding your tensor manipulation directly, what if you could just demonstrate it through …
inputs = {
'rows': [10, 20, 30],
'cols': [1, 2, 3, 4],
}
The desired output tensor, corresponding to the provided input tensors:output = [[11, 12, 13, 14],
[21, 22, 23, 24],
[31, 32, 33, 34]]
Given this information (already entered into the TF-Coder Colab by default), the TF-Coder tool will find the appropriate TensorFlow code automatically in a fraction of a second:tf.add(cols, tf.expand_dims(rows, 1))
The above problem was pretty simple just to illustrate the idea of programming by example. TF-Coder can be useful for harder problems as well, as we’ll see below.[10, 50, 100, 1000]
means that prices under $10 should fall into bucket 0, prices between $10 and $50 fall into bucket 1, and so on.# Input tensors
boundaries = [10, 50, 100, 1000]
prices = [15, 3, 50, 90, 100, 1001]
you want to compute the bucket number for each item:# Output tensor
bucketed_prices = [1, 0, 2, 2, 3, 4]
Although TensorFlow comes with various bucketing operations, it may be tricky to figure out which specific operation does this exact kind of bucketing. Since TF-Coder can identify hundreds of Tensor operations by behavior, you can look up the correct operation by providing an input-output example:# Input-output example
inputs = {
'boundaries': [10, 50, 100, 1000],
'prices': [15, 3, 50, 90, 100, 1001],
}
output = [1, 0, 2, 2, 3, 4]
Within seconds, TF-Coder outputs the following solution:tf.searchsorted(boundaries, prices, side='right')
This gives us a useful hint, and the documentation for tf.searchsorted
confirms that this code indeed performs the bucketing as desired.# Input tensor
scores = [[0.7, 0.2, 0.1],
[0.4, 0.5, 0.1],
[0.4, 0.4, 0.2],
[0.3, 0.4, 0.3],
[0.0, 0.0, 1.0]]
# Output tensor
top_scores = [[1, 0, 0],
[0, 1, 0],
[1, 0, 0],
[0, 1, 0],
[0, 0, 1]]
Note that if the same largest element appears multiple times within a row, such as in the third row of scores
, then only the first such largest element should be marked, so that every row of top_scores
has exactly one entry of 1
.tf.reduce_max
, tf.argmax
, and tf.maximum
are relevant, but which one should you use? tf.reduce_max
produces [0.7, 0.5, 0.4, 0.4, 1.0]
, tf.argmax
produces [0, 1, 0, 1, 2]
, and tf.maximum
isn’t right because it takes two arguments. None of these look close to our desired output.# Input-output example
inputs = {
'scores': [[0.7, 0.2, 0.1],
[0.4, 0.5, 0.1],
[0.4, 0.4, 0.2],
[0.3, 0.4, 0.3],
[0.0, 0.0, 1.0]],
}
output = [[1, 0, 0],
[0, 1, 0],
[1, 0, 0],
[0, 1, 0],
[0, 0, 1]]
TF-Coder uses a combination of tf.one_hot
and tf.argmax
in a short solution to this problem:tf.cast(tf.one_hot(tf.argmax(scores, axis=1), 3), tf.int32)
Through a detailed search over combinations of TensorFlow operations, TF-Coder often finds elegant solutions like this, which may simplify and speed up your TensorFlow programs.# Input tensor
counts = [[0, 1, 0, 0],
[0, 1, 1, 0],
[1, 1, 1, 1]]
# Output tensor
normalized = [[0.0, 1.0, 0.0, 0.0],
[0.0, 0.5, 0.5, 0.0],
[0.25, 0.25, 0.25, 0.25]]
Even if you know relevant functions to use (tf.reduce_sum
followed by tf.divide
), writing the correct code is still nontrivial. A first attempt may look like this:# First attempt
normalized = tf.divide(counts, tf.reduce_sum(counts, axis=1))
Is this right? There are many potential pitfalls to think about:axis=0
?counts
and tf.reduce_sum(counts, axis=1)
compatible for division, or do you need to reshape or transpose either of these?counts
and tf.reduce_sum(counts, axis=1)
are both tf.int32
tensors. Can tf.int32
tensors be divided, or do you need to cast them to a float DType first?tf.int32
, tf.float32
, or something else?# Input-output example
inputs = {
'counts': [[0, 1, 0, 0],
[0, 1, 1, 0],
[1, 1, 1, 1]],
}
output = [[0.0, 1.0, 0.0, 0.0],
[0.0, 0.5, 0.5, 0.0],
[0.25, 0.25, 0.25, 0.25]]
TF-Coder’s solution is: tf.cast(tf.divide(counts, tf.expand_dims(tf.reduce_sum(counts, axis=1), axis=1)), tf.float32)
By using TF-Coder to solve this problem, the mental burden of the exercise is reduced. When TF-Coder produces the solution above, it is guaranteed that the code correctly produces the example output when run on the example input. TF-Coder’s solution will also avoid any unnecessary steps. Thus, you can quickly deduce the answers to most of the questions above: an extra tf.expand_dims
step is needed to make the shapes compatible for division, and the result of tf.divide
must be cast to tf.float32
(in fact tf.divide
returns a tf.float64
tensor when dividing two tf.int32
tensors). In this way, TF-Coder helps you write simple and correct code without painful debugging cycles.
kolovoza 26, 2020
—
Posted by Kensen Shi, Google Research
When manipulating tensors, one must keep track of multiple dimensions, tensor shape and DType compatibility, and of course mathematical correctness. Additionally, there are hundreds of TensorFlow operations, and finding the right ones to use can be a challenge.
Instead of coding your tensor manipulation directly, what if you could just demonstrate it through …