[TOC]
##Ops for evaluation metrics and summary statistics.
This module provides functions for computing streaming metrics: metrics computed
on dynamically valued Tensors
. Each metric declaration returns a
"value_tensor", an idempotent operation that returns the current value of the
metric, and an "update_op", an operation that accumulates the information
from the current value of the Tensors
being measured as well as returns the
value of the "value_tensor".
To use any of these metrics, one need only declare the metric, call update_op
repeatedly to accumulate data over the desired number of Tensor
values (often
each one is a single batch) and finally evaluate the value_tensor. For example,
to use the streaming_mean
:
value = ...
mean_value, update_op = tf.contrib.metrics.streaming_mean(values)
sess.run(tf.local_variables_initializer())
for i in range(number_of_batches):
print('Mean after batch %d: %f' % (i, update_op.eval())
print('Final Mean: %f' % mean_value.eval())
Each metric function adds nodes to the graph that hold the state necessary to compute the value of the metric as well as a set of operations that actually perform the computation. Every metric evaluation is composed of three steps
- Initialization: initializing the metric state.
- Aggregation: updating the values of the metric state.
- Finalization: computing the final metric value.
In the above example, calling streaming_mean creates a pair of state variables
that will contain (1) the running sum and (2) the count of the number of samples
in the sum. Because the streaming metrics use local variables,
the Initialization stage is performed by running the op returned
by tf.local_variables_initializer()
. It sets the sum and count variables to
zero.
Next, Aggregation is performed by examining the current state of values
and incrementing the state variables appropriately. This step is executed by
running the update_op
returned by the metric.
Finally, finalization is performed by evaluating the "value_tensor"
In practice, we commonly want to evaluate across many batches and multiple metrics. To do so, we need only run the metric computation operations multiple times:
labels = ...
predictions = ...
accuracy, update_op_acc = tf.contrib.metrics.streaming_accuracy(
labels, predictions)
error, update_op_error = tf.contrib.metrics.streaming_mean_absolute_error(
labels, predictions)
sess.run(tf.local_variables_initializer())
for batch in range(num_batches):
sess.run([update_op_acc, update_op_error])
accuracy, mean_absolute_error = sess.run([accuracy, mean_absolute_error])
Note that when evaluating the same metric multiple times on different inputs, one must specify the scope of each metric to avoid accumulating the results together:
labels = ...
predictions0 = ...
predictions1 = ...
accuracy0 = tf.contrib.metrics.accuracy(labels, predictions0, name='preds0')
accuracy1 = tf.contrib.metrics.accuracy(labels, predictions1, name='preds1')
Certain metrics, such as streaming_mean or streaming_accuracy, can be weighted
via a weights
argument. The weights
tensor must be the same size as the
labels and predictions tensors and results in a weighted average of the metric.
tf.contrib.metrics.streaming_accuracy(predictions, labels, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_accuracy}
Calculates how often predictions
matches labels
.
The streaming_accuracy
function creates two local variables, total
and
count
that are used to compute the frequency with which predictions
matches labels
. This frequency is ultimately returned as accuracy
: an
idempotent operation that simply divides total
by count
.
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the accuracy
.
Internally, an is_correct
operation computes a Tensor
with elements 1.0
where the corresponding elements of predictions
and labels
match and 0.0
otherwise. Then update_op
increments total
with the reduced sum of the
product of weights
and is_correct
, and it increments count
with the
reduced sum of weights
.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: The predicted values, aTensor
of any shape.labels
: The ground truth values, aTensor
whose shape matchespredictions
.weights
:Tensor
whose rank is either 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections thataccuracy
should be added to.updates_collections
: An optional list of collections thatupdate_op
should be added to.name
: An optional variable_scope name.
accuracy
: ATensor
representing the accuracy, the value oftotal
divided bycount
.update_op
: An operation that increments thetotal
andcount
variables appropriately and whose value matchesaccuracy
.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, or ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_mean(values, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_mean}
Computes the (weighted) mean of the given values.
The streaming_mean
function creates two local variables, total
and count
that are used to compute the average of values
. This average is ultimately
returned as mean
which is an idempotent operation that simply divides
total
by count
.
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the mean
.
update_op
increments total
with the reduced sum of the product of values
and weights
, and it increments count
with the reduced sum of weights
.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
values
: ATensor
of arbitrary dimensions.weights
:Tensor
whose rank is either 0, or the same rank asvalues
, and must be broadcastable tovalues
(i.e., all dimensions must be either1
, or the same as the correspondingvalues
dimension).metrics_collections
: An optional list of collections thatmean
should be added to.updates_collections
: An optional list of collections thatupdate_op
should be added to.name
: An optional variable_scope name.
mean
: ATensor
representing the current mean, the value oftotal
divided bycount
.update_op
: An operation that increments thetotal
andcount
variables appropriately and whose value matchesmean_value
.
ValueError
: Ifweights
is notNone
and its shape doesn't matchvalues
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_recall(predictions, labels, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_recall}
Computes the recall of the predictions with respect to the labels.
The streaming_recall
function creates two local variables, true_positives
and false_negatives
, that are used to compute the recall. This value is
ultimately returned as recall
, an idempotent operation that simply divides
true_positives
by the sum of true_positives
and false_negatives
.
For estimation of the metric over a stream of data, the function creates an
update_op
that updates these variables and returns the recall
. update_op
weights each prediction by the corresponding value in weights
.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: The predicted values, abool
Tensor
of arbitrary shape.labels
: The ground truth values, abool
Tensor
whose dimensions must matchpredictions
.weights
:Tensor
whose rank is either 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections thatrecall
should be added to.updates_collections
: An optional list of collections thatupdate_op
should be added to.name
: An optional variable_scope name.
recall
: Scalar floatTensor
with the value oftrue_positives
divided by the sum oftrue_positives
andfalse_negatives
.update_op
:Operation
that incrementstrue_positives
andfalse_negatives
variables appropriately and whose value matchesrecall
.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, or ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_recall_at_thresholds(predictions, labels, thresholds, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_recall_at_thresholds}
Computes various recall values for different thresholds
on predictions
.
The streaming_recall_at_thresholds
function creates four local variables,
true_positives
, true_negatives
, false_positives
and false_negatives
for various values of thresholds. recall[i]
is defined as the total weight
of values in predictions
above thresholds[i]
whose corresponding entry in
labels
is True
, divided by the total weight of True
values in labels
(true_positives[i] / (true_positives[i] + false_negatives[i])
).
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the recall
.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: A floating pointTensor
of arbitrary shape and whose values are in the range[0, 1]
.labels
: Abool
Tensor
whose shape matchespredictions
.thresholds
: A python list or tuple of float thresholds in[0, 1]
.weights
:Tensor
whose rank is either 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections thatrecall
should be added to.updates_collections
: An optional list of collections thatupdate_op
should be added to.name
: An optional variable_scope name.
recall
: A floatTensor
of shape[len(thresholds)]
.update_op
: An operation that increments thetrue_positives
,true_negatives
,false_positives
andfalse_negatives
variables that are used in the computation ofrecall
.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, or ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_precision(predictions, labels, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_precision}
Computes the precision of the predictions with respect to the labels.
The streaming_precision
function creates two local variables,
true_positives
and false_positives
, that are used to compute the
precision. This value is ultimately returned as precision
, an idempotent
operation that simply divides true_positives
by the sum of true_positives
and false_positives
.
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the
precision
. update_op
weights each prediction by the corresponding value in
weights
.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: The predicted values, abool
Tensor
of arbitrary shape.labels
: The ground truth values, abool
Tensor
whose dimensions must matchpredictions
.weights
:Tensor
whose rank is either 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections thatprecision
should be added to.updates_collections
: An optional list of collections thatupdate_op
should be added to.name
: An optional variable_scope name.
precision
: Scalar floatTensor
with the value oftrue_positives
divided by the sum oftrue_positives
andfalse_positives
.update_op
:Operation
that incrementstrue_positives
andfalse_positives
variables appropriately and whose value matchesprecision
.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, or ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_precision_at_thresholds(predictions, labels, thresholds, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_precision_at_thresholds}
Computes precision values for different thresholds
on predictions
.
The streaming_precision_at_thresholds
function creates four local variables,
true_positives
, true_negatives
, false_positives
and false_negatives
for various values of thresholds. precision[i]
is defined as the total
weight of values in predictions
above thresholds[i]
whose corresponding
entry in labels
is True
, divided by the total weight of values in
predictions
above thresholds[i]
(true_positives[i] / (true_positives[i] + false_positives[i])
).
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the
precision
.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: A floating pointTensor
of arbitrary shape and whose values are in the range[0, 1]
.labels
: Abool
Tensor
whose shape matchespredictions
.thresholds
: A python list or tuple of float thresholds in[0, 1]
.weights
:Tensor
whose rank is either 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections thatauc
should be added to.updates_collections
: An optional list of collections thatupdate_op
should be added to.name
: An optional variable_scope name.
precision
: A floatTensor
of shape[len(thresholds)]
.update_op
: An operation that increments thetrue_positives
,true_negatives
,false_positives
andfalse_negatives
variables that are used in the computation ofprecision
.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, or ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_auc(predictions, labels, weights=None, num_thresholds=200, metrics_collections=None, updates_collections=None, curve='ROC', name=None)
{#streaming_auc}
Computes the approximate AUC via a Riemann sum.
The streaming_auc
function creates four local variables, true_positives
,
true_negatives
, false_positives
and false_negatives
that are used to
compute the AUC. To discretize the AUC curve, a linearly spaced set of
thresholds is used to compute pairs of recall and precision values. The area
under the ROC-curve is therefore computed using the height of the recall
values by the false positive rate, while the area under the PR-curve is the
computed using the height of the precision values by the recall.
This value is ultimately returned as auc
, an idempotent operation that
computes the area under a discretized curve of precision versus recall values
(computed using the aforementioned variables). The num_thresholds
variable
controls the degree of discretization with larger numbers of thresholds more
closely approximating the true AUC. The quality of the approximation may vary
dramatically depending on num_thresholds
.
For best results, predictions
should be distributed approximately uniformly
in the range [0, 1] and not peaked around 0 or 1. The quality of the AUC
approximation may be poor if this is not the case.
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the auc
.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
-
predictions
: A floating pointTensor
of arbitrary shape and whose values are in the range[0, 1]
. -
labels
: Abool
Tensor
whose shape matchespredictions
. -
weights
:Tensor
whose rank is either 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension). -
num_thresholds
: The number of thresholds to use when discretizing the roc curve. -
metrics_collections
: An optional list of collections thatauc
should be added to. -
updates_collections
: An optional list of collections thatupdate_op
should be added to. -
curve
: Specifies the name of the curve to be computed, 'ROC' [default] or 'PR' for the Precision-Recall-curve. -
name
: An optional variable_scope name.
auc
: A scalarTensor
representing the current area-under-curve.update_op
: An operation that increments thetrue_positives
,true_negatives
,false_positives
andfalse_negatives
variables appropriately and whose value matchesauc
.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, or ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
Computes the recall@k of the predictions with respect to dense labels. (deprecated)
THIS FUNCTION IS DEPRECATED. It will be removed after 2016-11-08.
Instructions for updating:
Please use streaming_sparse_recall_at_k
, and reshape labels from [batch_size] to [batch_size, 1].
The streaming_recall_at_k
function creates two local variables, total
and
count
, that are used to compute the recall@k frequency. This frequency is
ultimately returned as recall_at_<k>
: an idempotent operation that simply
divides total
by count
.
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the
recall_at_<k>
. Internally, an in_top_k
operation computes a Tensor
with
shape [batch_size] whose elements indicate whether or not the corresponding
label is in the top k
predictions
. Then update_op
increments total
with the reduced sum of weights
where in_top_k
is True
, and it
increments count
with the reduced sum of weights
.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: A floatTensor
of dimension [batch_size, num_classes].labels
: ATensor
of dimension [batch_size] whose type is inint32
,int64
.k
: The number of top elements to look at for computing recall.weights
:Tensor
whose rank is either 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections thatrecall_at_k
should be added to.updates_collections
: An optional list of collectionsupdate_op
should be added to.name
: An optional variable_scope name.
recall_at_k
: ATensor
representing the recall@k, the fraction of labels which fall into the topk
predictions.update_op
: An operation that increments thetotal
andcount
variables appropriately and whose value matchesrecall_at_k
.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, or ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_mean_absolute_error(predictions, labels, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_mean_absolute_error}
Computes the mean absolute error between the labels and predictions.
The streaming_mean_absolute_error
function creates two local variables,
total
and count
that are used to compute the mean absolute error. This
average is weighted by weights
, and it is ultimately returned as
mean_absolute_error
: an idempotent operation that simply divides total
by
count
.
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the
mean_absolute_error
. Internally, an absolute_errors
operation computes the
absolute value of the differences between predictions
and labels
. Then
update_op
increments total
with the reduced sum of the product of
weights
and absolute_errors
, and it increments count
with the reduced
sum of weights
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: ATensor
of arbitrary shape.labels
: ATensor
of the same shape aspredictions
.weights
: OptionalTensor
indicating the frequency with which an example is sampled. Rank must be 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections thatmean_absolute_error
should be added to.updates_collections
: An optional list of collections thatupdate_op
should be added to.name
: An optional variable_scope name.
mean_absolute_error
: ATensor
representing the current mean, the value oftotal
divided bycount
.update_op
: An operation that increments thetotal
andcount
variables appropriately and whose value matchesmean_absolute_error
.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, or ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_mean_iou(predictions, labels, num_classes, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_mean_iou}
Calculate per-step mean Intersection-Over-Union (mIOU).
Mean Intersection-Over-Union is a common evaluation metric for semantic image segmentation, which first computes the IOU for each semantic class and then computes the average over classes.
IOU = true_positive / (true_positive + false_positive + false_negative).
The predictions are accumulated in a confusion matrix, weighted by weights
,
and mIOU is then calculated from it.
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the mean_iou
.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: ATensor
of prediction results for semantic labels, whose shape is [batch size] and typeint32
orint64
. The tensor will be flattened, if its rank > 1.labels
: ATensor
of ground truth labels with shape [batch size] and of typeint32
orint64
. The tensor will be flattened, if its rank > 1.num_classes
: The possible number of labels the prediction task can have. This value must be provided, since a confusion matrix of dimension = [num_classes, num_classes] will be allocated.weights
: An optionalTensor
whose shape is broadcastable topredictions
.metrics_collections
: An optional list of collections thatmean_iou
should be added to.updates_collections
: An optional list of collectionsupdate_op
should be added to.name
: An optional variable_scope name.
mean_iou
: ATensor
representing the mean intersection-over-union.update_op
: An operation that increments the confusion matrix.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, or ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_mean_relative_error(predictions, labels, normalizer, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_mean_relative_error}
Computes the mean relative error by normalizing with the given values.
The streaming_mean_relative_error
function creates two local variables,
total
and count
that are used to compute the mean relative absolute error.
This average is weighted by weights
, and it is ultimately returned as
mean_relative_error
: an idempotent operation that simply divides total
by
count
.
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the
mean_reative_error
. Internally, a relative_errors
operation divides the
absolute value of the differences between predictions
and labels
by the
normalizer
. Then update_op
increments total
with the reduced sum of the
product of weights
and relative_errors
, and it increments count
with the
reduced sum of weights
.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: ATensor
of arbitrary shape.labels
: ATensor
of the same shape aspredictions
.normalizer
: ATensor
of the same shape aspredictions
.weights
: OptionalTensor
indicating the frequency with which an example is sampled. Rank must be 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections thatmean_relative_error
should be added to.updates_collections
: An optional list of collections thatupdate_op
should be added to.name
: An optional variable_scope name.
mean_relative_error
: ATensor
representing the current mean, the value oftotal
divided bycount
.update_op
: An operation that increments thetotal
andcount
variables appropriately and whose value matchesmean_relative_error
.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, or ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_mean_squared_error(predictions, labels, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_mean_squared_error}
Computes the mean squared error between the labels and predictions.
The streaming_mean_squared_error
function creates two local variables,
total
and count
that are used to compute the mean squared error.
This average is weighted by weights
, and it is ultimately returned as
mean_squared_error
: an idempotent operation that simply divides total
by
count
.
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the
mean_squared_error
. Internally, a squared_error
operation computes the
element-wise square of the difference between predictions
and labels
. Then
update_op
increments total
with the reduced sum of the product of
weights
and squared_error
, and it increments count
with the reduced sum
of weights
.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: ATensor
of arbitrary shape.labels
: ATensor
of the same shape aspredictions
.weights
: OptionalTensor
indicating the frequency with which an example is sampled. Rank must be 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections thatmean_squared_error
should be added to.updates_collections
: An optional list of collections thatupdate_op
should be added to.name
: An optional variable_scope name.
mean_squared_error
: ATensor
representing the current mean, the value oftotal
divided bycount
.update_op
: An operation that increments thetotal
andcount
variables appropriately and whose value matchesmean_squared_error
.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, or ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_root_mean_squared_error(predictions, labels, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_root_mean_squared_error}
Computes the root mean squared error between the labels and predictions.
The streaming_root_mean_squared_error
function creates two local variables,
total
and count
that are used to compute the root mean squared error.
This average is weighted by weights
, and it is ultimately returned as
root_mean_squared_error
: an idempotent operation that takes the square root
of the division of total
by count
.
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the
root_mean_squared_error
. Internally, a squared_error
operation computes
the element-wise square of the difference between predictions
and labels
.
Then update_op
increments total
with the reduced sum of the product of
weights
and squared_error
, and it increments count
with the reduced sum
of weights
.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: ATensor
of arbitrary shape.labels
: ATensor
of the same shape aspredictions
.weights
: OptionalTensor
indicating the frequency with which an example is sampled. Rank must be 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections thatroot_mean_squared_error
should be added to.updates_collections
: An optional list of collections thatupdate_op
should be added to.name
: An optional variable_scope name.
root_mean_squared_error
: ATensor
representing the current mean, the value oftotal
divided bycount
.update_op
: An operation that increments thetotal
andcount
variables appropriately and whose value matchesroot_mean_squared_error
.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, or ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_covariance(predictions, labels, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_covariance}
Computes the unbiased sample covariance between predictions
and labels
.
The streaming_covariance
function creates four local variables,
comoment
, mean_prediction
, mean_label
, and count
, which are used to
compute the sample covariance between predictions and labels across multiple
batches of data. The covariance is ultimately returned as an idempotent
operation that simply divides comoment
by count
- 1. We use count
- 1
in order to get an unbiased estimate.
The algorithm used for this online computation is described in
https://en.wikipedia.org/wiki/Algorithms_for_calculating_variance.
Specifically, the formula used to combine two sample comoments is
C_AB = C_A + C_B + (E[x_A] - E[x_B]) * (E[y_A] - E[y_B]) * n_A * n_B / n_AB
The comoment for a single batch of data is simply
sum((x - E[x]) * (y - E[y]))
, optionally weighted.
If weights
is not None, then it is used to compute weighted comoments,
means, and count. NOTE: these weights are treated as "frequency weights", as
opposed to "reliability weights". See discussion of the difference on
https://wikipedia.org/wiki/Weighted_arithmetic_mean#Weighted_sample_variance
To facilitate the computation of covariance across multiple batches of data,
the function creates an update_op
operation, which updates underlying
variables and returns the updated covariance.
predictions
: ATensor
of arbitrary size.labels
: ATensor
of the same size aspredictions
.weights
: OptionalTensor
indicating the frequency with which an example is sampled. Rank must be 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections that the metric value variable should be added to.updates_collections
: An optional list of collections that the metric update ops should be added to.name
: An optional variable_scope name.
covariance
: ATensor
representing the current unbiased sample covariance,comoment
/ (count
- 1).update_op
: An operation that updates the local variables appropriately.
ValueError
: If labels and predictions are of different sizes or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_pearson_correlation(predictions, labels, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_pearson_correlation}
Computes Pearson correlation coefficient between predictions
, labels
.
The streaming_pearson_correlation
function delegates to
streaming_covariance
the tracking of three [co]variances:
streaming_covariance(predictions, labels)
, i.e. covariancestreaming_covariance(predictions, predictions)
, i.e. variancestreaming_covariance(labels, labels)
, i.e. variance
The product-moment correlation ultimately returned is an idempotent operation
cov(predictions, labels) / sqrt(var(predictions) * var(labels))
. To
facilitate correlation computation across multiple batches, the function
groups the update_op
s of the underlying streaming_covariance and returns an
update_op
.
If weights
is not None, then it is used to compute a weighted correlation.
NOTE: these weights are treated as "frequency weights", as opposed to
"reliability weights". See discussion of the difference on
https://wikipedia.org/wiki/Weighted_arithmetic_mean#Weighted_sample_variance
predictions
: ATensor
of arbitrary size.labels
: ATensor
of the same size as predictions.weights
: OptionalTensor
indicating the frequency with which an example is sampled. Rank must be 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections that the metric value variable should be added to.updates_collections
: An optional list of collections that the metric update ops should be added to.name
: An optional variable_scope name.
pearson_r
: ATensor
representing the current Pearson product-moment correlation coefficient, the value ofcov(predictions, labels) / sqrt(var(predictions) * var(labels))
.update_op
: An operation that updates the underlying variables appropriately.
ValueError
: Iflabels
andpredictions
are of different sizes, or ifweights
is the wrong size, or if eithermetrics_collections
orupdates_collections
are not alist
ortuple
.
tf.contrib.metrics.streaming_mean_cosine_distance(predictions, labels, dim, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_mean_cosine_distance}
Computes the cosine distance between the labels and predictions.
The streaming_mean_cosine_distance
function creates two local variables,
total
and count
that are used to compute the average cosine distance
between predictions
and labels
. This average is weighted by weights
,
and it is ultimately returned as mean_distance
, which is an idempotent
operation that simply divides total
by count
.
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the
mean_distance
.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: ATensor
of the same shape aslabels
.labels
: ATensor
of arbitrary shape.dim
: The dimension along which the cosine distance is computed.weights
: An optionalTensor
whose shape is broadcastable topredictions
, and whose dimensiondim
is 1.metrics_collections
: An optional list of collections that the metric value variable should be added to.updates_collections
: An optional list of collections that the metric update ops should be added to.name
: An optional variable_scope name.
mean_distance
: ATensor
representing the current mean, the value oftotal
divided bycount
.update_op
: An operation that increments thetotal
andcount
variables appropriately.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, or ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_percentage_less(values, threshold, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_percentage_less}
Computes the percentage of values less than the given threshold.
The streaming_percentage_less
function creates two local variables,
total
and count
that are used to compute the percentage of values
that
fall below threshold
. This rate is weighted by weights
, and it is
ultimately returned as percentage
which is an idempotent operation that
simply divides total
by count
.
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the
percentage
.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
values
: A numericTensor
of arbitrary size.threshold
: A scalar threshold.weights
: An optionalTensor
whose shape is broadcastable tovalues
.metrics_collections
: An optional list of collections that the metric value variable should be added to.updates_collections
: An optional list of collections that the metric update ops should be added to.name
: An optional variable_scope name.
percentage
: ATensor
representing the current mean, the value oftotal
divided bycount
.update_op
: An operation that increments thetotal
andcount
variables appropriately.
ValueError
: Ifweights
is notNone
and its shape doesn't matchvalues
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_sensitivity_at_specificity(predictions, labels, specificity, weights=None, num_thresholds=200, metrics_collections=None, updates_collections=None, name=None)
{#streaming_sensitivity_at_specificity}
Computes the specificity at a given sensitivity.
The streaming_sensitivity_at_specificity
function creates four local
variables, true_positives
, true_negatives
, false_positives
and
false_negatives
that are used to compute the sensitivity at the given
specificity value. The threshold for the given specificity value is computed
and used to evaluate the corresponding sensitivity.
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the
sensitivity
. update_op
increments the true_positives
, true_negatives
,
false_positives
and false_negatives
counts with the weight of each case
found in the predictions
and labels
.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
For additional information about specificity and sensitivity, see the following: https://en.wikipedia.org/wiki/Sensitivity_and_specificity
predictions
: A floating pointTensor
of arbitrary shape and whose values are in the range[0, 1]
.labels
: Abool
Tensor
whose shape matchespredictions
.specificity
: A scalar value in range[0, 1]
.weights
:Tensor
whose rank is either 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).num_thresholds
: The number of thresholds to use for matching the given specificity.metrics_collections
: An optional list of collections thatsensitivity
should be added to.updates_collections
: An optional list of collections thatupdate_op
should be added to.name
: An optional variable_scope name.
sensitivity
: A scalarTensor
representing the sensitivity at the givenspecificity
value.update_op
: An operation that increments thetrue_positives
,true_negatives
,false_positives
andfalse_negatives
variables appropriately and whose value matchessensitivity
.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, ifweights
is notNone
and its shape doesn't matchpredictions
, or ifspecificity
is not between 0 and 1, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_sparse_average_precision_at_k(predictions, labels, k, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_sparse_average_precision_at_k}
Computes average precision@k of predictions with respect to sparse labels.
See sparse_average_precision_at_k
for details on formula. weights
are
applied to the result of sparse_average_precision_at_k
streaming_sparse_average_precision_at_k
creates two local variables,
average_precision_at_<k>/total
and average_precision_at_<k>/max
, that
are used to compute the frequency. This frequency is ultimately returned as
average_precision_at_<k>
: an idempotent operation that simply divides
average_precision_at_<k>/total
by average_precision_at_<k>/max
.
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the
precision_at_<k>
. Internally, a top_k
operation computes a Tensor
indicating the top k
predictions
. Set operations applied to top_k
and
labels
calculate the true positives and false positives weighted by
weights
. Then update_op
increments true_positive_at_<k>
and
false_positive_at_<k>
using these values.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: FloatTensor
with shape [D1, ... DN, num_classes] where N >= 1. Commonly, N=1 andpredictions
has shape [batch size, num_classes]. The final dimension contains the logit values for each class. [D1, ... DN] must matchlabels
.labels
:int64
Tensor
orSparseTensor
with shape [D1, ... DN, num_labels], where N >= 1 and num_labels is the number of target classes for the associated prediction. Commonly, N=1 andlabels
has shape [batch_size, num_labels]. [D1, ... DN] must matchpredictions_
. Values should be in range [0, num_classes), where num_classes is the last dimension ofpredictions
. Values outside this range are ignored.k
: Integer, k for @k metric. This will calculate an average precision for range[1,k]
, as documented above.weights
:Tensor
whose rank is either 0, or n-1, where n is the rank oflabels
. If the latter, it must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections that values should be added to.updates_collections
: An optional list of collections that updates should be added to.name
: Name of new update operation, and namespace for other dependent ops.
mean_average_precision
: Scalarfloat64
Tensor
with the mean average precision values.update
:Operation
that increments variables appropriately, and whose value matchesmetric
.
tf.contrib.metrics.streaming_sparse_precision_at_k(predictions, labels, k, class_id=None, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_sparse_precision_at_k}
Computes precision@k of the predictions with respect to sparse labels.
If class_id
is not specified, we calculate precision as the ratio of true
positives (i.e., correct predictions, items in the top k
highest
predictions
that are found in the corresponding row in labels
) to
positives (all top k
predictions
).
If class_id
is specified, we calculate precision by considering only the
rows in the batch for which class_id
is in the top k
highest
predictions
, and computing the fraction of them for which class_id
is
in the corresponding row in labels
.
We expect precision to decrease as k
increases.
streaming_sparse_precision_at_k
creates two local variables,
true_positive_at_<k>
and false_positive_at_<k>
, that are used to compute
the precision@k frequency. This frequency is ultimately returned as
precision_at_<k>
: an idempotent operation that simply divides
true_positive_at_<k>
by total (true_positive_at_<k>
+
false_positive_at_<k>
).
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the
precision_at_<k>
. Internally, a top_k
operation computes a Tensor
indicating the top k
predictions
. Set operations applied to top_k
and
labels
calculate the true positives and false positives weighted by
weights
. Then update_op
increments true_positive_at_<k>
and
false_positive_at_<k>
using these values.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: FloatTensor
with shape [D1, ... DN, num_classes] where N >= 1. Commonly, N=1 and predictions has shape [batch size, num_classes]. The final dimension contains the logit values for each class. [D1, ... DN] must matchlabels
.labels
:int64
Tensor
orSparseTensor
with shape [D1, ... DN, num_labels], where N >= 1 and num_labels is the number of target classes for the associated prediction. Commonly, N=1 andlabels
has shape [batch_size, num_labels]. [D1, ... DN] must matchpredictions
. Values should be in range [0, num_classes), where num_classes is the last dimension ofpredictions
. Values outside this range are ignored.k
: Integer, k for @k metric.class_id
: Integer class ID for which we want binary metrics. This should be in range [0, num_classes], where num_classes is the last dimension ofpredictions
. Ifclass_id
is outside this range, the method returns NAN.weights
:Tensor
whose rank is either 0, or n-1, where n is the rank oflabels
. If the latter, it must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections that values should be added to.updates_collections
: An optional list of collections that updates should be added to.name
: Name of new update operation, and namespace for other dependent ops.
precision
: Scalarfloat64
Tensor
with the value oftrue_positives
divided by the sum oftrue_positives
andfalse_positives
.update_op
:Operation
that incrementstrue_positives
andfalse_positives
variables appropriately, and whose value matchesprecision
.
ValueError
: Ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_sparse_precision_at_top_k(top_k_predictions, labels, class_id=None, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_sparse_precision_at_top_k}
Computes precision@k of top-k predictions with respect to sparse labels.
If class_id
is not specified, we calculate precision as the ratio of
true positives (i.e., correct predictions, items in top_k_predictions
that are found in the corresponding row in labels
) to positives (all
top_k_predictions
).
If class_id
is specified, we calculate precision by considering only the
rows in the batch for which class_id
is in the top k
highest
predictions
, and computing the fraction of them for which class_id
is
in the corresponding row in labels
.
We expect precision to decrease as k
increases.
streaming_sparse_precision_at_top_k
creates two local variables,
true_positive_at_k
and false_positive_at_k
, that are used to compute
the precision@k frequency. This frequency is ultimately returned as
precision_at_k
: an idempotent operation that simply divides
true_positive_at_k
by total (true_positive_at_k
+ false_positive_at_k
).
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the
precision_at_k
. Internally, set operations applied to top_k_predictions
and labels
calculate the true positives and false positives weighted by
weights
. Then update_op
increments true_positive_at_k
and
false_positive_at_k
using these values.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
top_k_predictions
: IntegerTensor
with shape [D1, ... DN, k] where N >= 1. Commonly, N=1 and top_k_predictions has shape [batch size, k]. The final dimension contains the indices of top-k labels. [D1, ... DN] must matchlabels
.labels
:int64
Tensor
orSparseTensor
with shape [D1, ... DN, num_labels], where N >= 1 and num_labels is the number of target classes for the associated prediction. Commonly, N=1 andlabels
has shape [batch_size, num_labels]. [D1, ... DN] must matchtop_k_predictions
. Values should be in range [0, num_classes), where num_classes is the last dimension ofpredictions
. Values outside this range are ignored.class_id
: Integer class ID for which we want binary metrics. This should be in range [0, num_classes), where num_classes is the last dimension ofpredictions
. Ifclass_id
is outside this range, the method returns NAN.weights
:Tensor
whose rank is either 0, or n-1, where n is the rank oflabels
. If the latter, it must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections that values should be added to.updates_collections
: An optional list of collections that updates should be added to.name
: Name of new update operation, and namespace for other dependent ops.
precision
: Scalarfloat64
Tensor
with the value oftrue_positives
divided by the sum oftrue_positives
andfalse_positives
.update_op
:Operation
that incrementstrue_positives
andfalse_positives
variables appropriately, and whose value matchesprecision
.
ValueError
: Ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.ValueError
: Iftop_k_predictions
has rank < 2.
tf.contrib.metrics.streaming_sparse_recall_at_k(predictions, labels, k, class_id=None, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_sparse_recall_at_k}
Computes recall@k of the predictions with respect to sparse labels.
If class_id
is not specified, we'll calculate recall as the ratio of true
positives (i.e., correct predictions, items in the top k
highest
predictions
that are found in the corresponding row in labels
) to
actual positives (the full labels
row).
If class_id
is specified, we calculate recall by considering only the rows
in the batch for which class_id
is in labels
, and computing the
fraction of them for which class_id
is in the corresponding row in
labels
.
streaming_sparse_recall_at_k
creates two local variables,
true_positive_at_<k>
and false_negative_at_<k>
, that are used to compute
the recall_at_k frequency. This frequency is ultimately returned as
recall_at_<k>
: an idempotent operation that simply divides
true_positive_at_<k>
by total (true_positive_at_<k>
+
false_negative_at_<k>
).
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the
recall_at_<k>
. Internally, a top_k
operation computes a Tensor
indicating the top k
predictions
. Set operations applied to top_k
and
labels
calculate the true positives and false negatives weighted by
weights
. Then update_op
increments true_positive_at_<k>
and
false_negative_at_<k>
using these values.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: FloatTensor
with shape [D1, ... DN, num_classes] where N >= 1. Commonly, N=1 and predictions has shape [batch size, num_classes]. The final dimension contains the logit values for each class. [D1, ... DN] must matchlabels
.labels
:int64
Tensor
orSparseTensor
with shape [D1, ... DN, num_labels], where N >= 1 and num_labels is the number of target classes for the associated prediction. Commonly, N=1 andlabels
has shape [batch_size, num_labels]. [D1, ... DN] must matchpredictions
. Values should be in range [0, num_classes), where num_classes is the last dimension ofpredictions
. Values outside this range always count towardsfalse_negative_at_<k>
.k
: Integer, k for @k metric.class_id
: Integer class ID for which we want binary metrics. This should be in range [0, num_classes), where num_classes is the last dimension ofpredictions
. If class_id is outside this range, the method returns NAN.weights
:Tensor
whose rank is either 0, or n-1, where n is the rank oflabels
. If the latter, it must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections that values should be added to.updates_collections
: An optional list of collections that updates should be added to.name
: Name of new update operation, and namespace for other dependent ops.
recall
: Scalarfloat64
Tensor
with the value oftrue_positives
divided by the sum oftrue_positives
andfalse_negatives
.update_op
:Operation
that incrementstrue_positives
andfalse_negatives
variables appropriately, and whose value matchesrecall
.
ValueError
: Ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_specificity_at_sensitivity(predictions, labels, sensitivity, weights=None, num_thresholds=200, metrics_collections=None, updates_collections=None, name=None)
{#streaming_specificity_at_sensitivity}
Computes the specificity at a given sensitivity.
The streaming_specificity_at_sensitivity
function creates four local
variables, true_positives
, true_negatives
, false_positives
and
false_negatives
that are used to compute the specificity at the given
sensitivity value. The threshold for the given sensitivity value is computed
and used to evaluate the corresponding specificity.
For estimation of the metric over a stream of data, the function creates an
update_op
operation that updates these variables and returns the
specificity
. update_op
increments the true_positives
, true_negatives
,
false_positives
and false_negatives
counts with the weight of each case
found in the predictions
and labels
.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
For additional information about specificity and sensitivity, see the following: https://en.wikipedia.org/wiki/Sensitivity_and_specificity
predictions
: A floating pointTensor
of arbitrary shape and whose values are in the range[0, 1]
.labels
: Abool
Tensor
whose shape matchespredictions
.sensitivity
: A scalar value in range[0, 1]
.weights
:Tensor
whose rank is either 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).num_thresholds
: The number of thresholds to use for matching the given sensitivity.metrics_collections
: An optional list of collections thatspecificity
should be added to.updates_collections
: An optional list of collections thatupdate_op
should be added to.name
: An optional variable_scope name.
specificity
: A scalarTensor
representing the specificity at the givenspecificity
value.update_op
: An operation that increments thetrue_positives
,true_negatives
,false_positives
andfalse_negatives
variables appropriately and whose value matchesspecificity
.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, ifweights
is notNone
and its shape doesn't matchpredictions
, or ifsensitivity
is not between 0 and 1, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_concat(values, axis=0, max_size=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_concat}
Concatenate values along an axis across batches.
The function streaming_concat
creates two local variables, array
and
size
, that are used to store concatenated values. Internally, array
is
used as storage for a dynamic array (if maxsize
is None
), which ensures
that updates can be run in amortized constant time.
For estimation of the metric over a stream of data, the function creates an
update_op
operation that appends the values of a tensor and returns the
value
of the concatenated tensors.
This op allows for evaluating metrics that cannot be updated incrementally using the same framework as other streaming metrics.
values
:Tensor
to concatenate. Rank and the shape along all axes other than the axis to concatenate along must be statically known.axis
: optional integer axis to concatenate along.max_size
: optional integer maximum size ofvalue
along the given axis. Once the maximum size is reached, further updates are no-ops. By default, there is no maximum size: the array is resized as necessary.metrics_collections
: An optional list of collections thatvalue
should be added to.updates_collections
: An optional list of collectionsupdate_op
should be added to.name
: An optional variable_scope name.
value
: ATensor
representing the concatenated values.update_op
: An operation that concatenates the next values.
ValueError
: ifvalues
does not have a statically known rank,axis
is not in the valid range or the size ofvalues
is not statically known along any axis other thanaxis
.
tf.contrib.metrics.streaming_false_negatives(predictions, labels, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_false_negatives}
Computes the total number of false positives.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: The predicted values, abool
Tensor
of arbitrary dimensions.labels
: The ground truth values, abool
Tensor
whose dimensions must matchpredictions
.weights
: OptionalTensor
whose rank is either 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections that the metric value variable should be added to.updates_collections
: An optional list of collections that the metric update ops should be added to.name
: An optional variable_scope name.
value_tensor
: ATensor
representing the current value of the metric.update_op
: An operation that accumulates the error from a batch of data.
ValueError
: Ifweights
is notNone
and its shape doesn't matchvalues
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_false_negatives_at_thresholds(predictions, labels, thresholds, weights=None)
{#streaming_false_negatives_at_thresholds}
tf.contrib.metrics.streaming_false_positives(predictions, labels, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_false_positives}
Sum the weights of false positives.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: The predicted values, abool
Tensor
of arbitrary dimensions.labels
: The ground truth values, abool
Tensor
whose dimensions must matchpredictions
.weights
: OptionalTensor
whose rank is either 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections that the metric value variable should be added to.updates_collections
: An optional list of collections that the metric update ops should be added to.name
: An optional variable_scope name.
value_tensor
: ATensor
representing the current value of the metric.update_op
: An operation that accumulates the error from a batch of data.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, or ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_false_positives_at_thresholds(predictions, labels, thresholds, weights=None)
{#streaming_false_positives_at_thresholds}
tf.contrib.metrics.streaming_true_negatives(predictions, labels, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_true_negatives}
Sum the weights of true_negatives.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: The predicted values, abool
Tensor
of arbitrary dimensions.labels
: The ground truth values, abool
Tensor
whose dimensions must matchpredictions
.weights
: OptionalTensor
whose rank is either 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections that the metric value variable should be added to.updates_collections
: An optional list of collections that the metric update ops should be added to.name
: An optional variable_scope name.
value_tensor
: ATensor
representing the current value of the metric.update_op
: An operation that accumulates the error from a batch of data.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, or ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_true_negatives_at_thresholds(predictions, labels, thresholds, weights=None)
{#streaming_true_negatives_at_thresholds}
tf.contrib.metrics.streaming_true_positives(predictions, labels, weights=None, metrics_collections=None, updates_collections=None, name=None)
{#streaming_true_positives}
Sum the weights of true_positives.
If weights
is None
, weights default to 1. Use weights of 0 to mask values.
predictions
: The predicted values, abool
Tensor
of arbitrary dimensions.labels
: The ground truth values, abool
Tensor
whose dimensions must matchpredictions
.weights
: OptionalTensor
whose rank is either 0, or the same rank aslabels
, and must be broadcastable tolabels
(i.e., all dimensions must be either1
, or the same as the correspondinglabels
dimension).metrics_collections
: An optional list of collections that the metric value variable should be added to.updates_collections
: An optional list of collections that the metric update ops should be added to.name
: An optional variable_scope name.
value_tensor
: ATensor
representing the current value of the metric.update_op
: An operation that accumulates the error from a batch of data.
ValueError
: Ifpredictions
andlabels
have mismatched shapes, or ifweights
is notNone
and its shape doesn't matchpredictions
, or if eithermetrics_collections
orupdates_collections
are not a list or tuple.
tf.contrib.metrics.streaming_true_positives_at_thresholds(predictions, labels, thresholds, weights=None)
{#streaming_true_positives_at_thresholds}
tf.contrib.metrics.auc_using_histogram(boolean_labels, scores, score_range, nbins=100, collections=None, check_shape=True, name=None)
{#auc_using_histogram}
AUC computed by maintaining histograms.
Rather than computing AUC directly, this Op maintains Variables containing
histograms of the scores associated with True
and False
labels. By
comparing these the AUC is generated, with some discretization error.
See: "Efficient AUC Learning Curve Calculation" by Bouckaert.
This AUC Op updates in O(batch_size + nbins)
time and works well even with
large class imbalance. The accuracy is limited by discretization error due
to finite number of bins. If scores are concentrated in a fewer bins,
accuracy is lower. If this is a concern, we recommend trying different
numbers of bins and comparing results.
boolean_labels
: 1-D booleanTensor
. Entry isTrue
if the corresponding record is in class.scores
: 1-D numericTensor
, same shape as boolean_labels.score_range
:Tensor
of shape[2]
, same dtype asscores
. The min/max values of score that we expect. Scores outside range will be clipped.nbins
: Integer number of bins to use. Accuracy strictly increases as the number of bins increases.collections
: List of graph collections keys. Internal histogram Variables are added to these collections. Defaults to[GraphKeys.LOCAL_VARIABLES]
.check_shape
: Boolean. IfTrue
, do a runtime shape check on the scores and labels.name
: A name for this Op. Defaults to "auc_using_histogram".
auc
:float32
scalarTensor
. Fetching this converts internal histograms to auc value.update_op
:Op
, when run, updates internal histograms.
Computes the percentage of times that predictions matches labels.
predictions
: the predicted values, aTensor
whose dtype and shape matches 'labels'.labels
: the ground truth values, aTensor
of any shape and bool, integer, or string dtype.weights
: None orTensor
of float values to reweight the accuracy.
Accuracy Tensor
.
ValueError
: if dtypes don't match or if dtype is not bool, integer, or string.
Aggregates the metric value tensors and update ops into two lists.
*value_update_tuples
: a variable number of tuples, each of which contain the pair of (value_tensor, update_op) from a streaming metric.
A list of value Tensor
objects and a list of update ops.
ValueError
: ifvalue_update_tuples
is empty.
Aggregates the metric names to tuple dictionary.
This function is useful for pairing metric names with their associated value and update ops when the list of metrics is long. For example:
metrics_to_values, metrics_to_updates = slim.metrics.aggregate_metric_map({
'Mean Absolute Error': new_slim.metrics.streaming_mean_absolute_error(
predictions, labels, weights),
'Mean Relative Error': new_slim.metrics.streaming_mean_relative_error(
predictions, labels, labels, weights),
'RMSE Linear': new_slim.metrics.streaming_root_mean_squared_error(
predictions, labels, weights),
'RMSE Log': new_slim.metrics.streaming_root_mean_squared_error(
predictions, labels, weights),
})
names_to_tuples
: a map of metric names to tuples, each of which contain the pair of (value_tensor, update_op) from a streaming metric.
A dictionary from metric names to value ops and a dictionary from metric names to update ops.
Compute set difference of elements in last dimension of a
and b
.
All but the last dimension of a
and b
must match.
Example:
a = [
[
[
[1, 2],
[3],
],
[
[4],
[5, 6],
],
],
]
b = [
[
[
[1, 3],
[2],
],
[
[4, 5],
[5, 6, 7, 8],
],
],
]
set_difference(a, b, aminusb=True) = [
[
[
[2],
[3],
],
[
[],
[],
],
],
]
a
:Tensor
orSparseTensor
of the same type asb
. If sparse, indices must be sorted in row-major order.b
:Tensor
orSparseTensor
of the same type asa
. If sparse, indices must be sorted in row-major order.aminusb
: Whether to subtractb
froma
, vs vice versa.validate_indices
: Whether to validate the order and range of sparse indices ina
andb
.
A SparseTensor
whose shape is the same rank as a
and b
, and all but
the last dimension the same. Elements along the last dimension contain the
differences.
Compute set intersection of elements in last dimension of a
and b
.
All but the last dimension of a
and b
must match.
Example:
a = [
[
[
[1, 2],
[3],
],
[
[4],
[5, 6],
],
],
]
b = [
[
[
[1, 3],
[2],
],
[
[4, 5],
[5, 6, 7, 8],
],
],
]
set_intersection(a, b) = [
[
[
[1],
[],
],
[
[4],
[5, 6],
],
],
]
a
:Tensor
orSparseTensor
of the same type asb
. If sparse, indices must be sorted in row-major order.b
:Tensor
orSparseTensor
of the same type asa
. If sparse, indices must be sorted in row-major order.validate_indices
: Whether to validate the order and range of sparse indices ina
andb
.
A SparseTensor
whose shape is the same rank as a
and b
, and all but
the last dimension the same. Elements along the last dimension contain the
intersections.
Compute number of unique elements along last dimension of a
.
a
:SparseTensor
, with indices sorted in row-major order.validate_indices
: Whether to validate the order and range of sparse indices ina
.
int32
Tensor
of set sizes. For a
ranked n
, this is a Tensor
with
rank n-1
, and the same 1st n-1
dimensions as a
. Each value is the
number of unique elements in the corresponding [0...n-1]
dimension of a
.
TypeError
: Ifa
is an invalid types.
Compute set union of elements in last dimension of a
and b
.
All but the last dimension of a
and b
must match.
Example:
a = [
[
[
[1, 2],
[3],
],
[
[4],
[5, 6],
],
],
]
b = [
[
[
[1, 3],
[2],
],
[
[4, 5],
[5, 6, 7, 8],
],
],
]
set_union(a, b) = [
[
[
[1, 2, 3],
[2, 3],
],
[
[4, 5],
[5, 6, 7, 8],
],
],
]
a
:Tensor
orSparseTensor
of the same type asb
. If sparse, indices must be sorted in row-major order.b
:Tensor
orSparseTensor
of the same type asa
. If sparse, indices must be sorted in row-major order.validate_indices
: Whether to validate the order and range of sparse indices ina
andb
.
A SparseTensor
whose shape is the same rank as a
and b
, and all but
the last dimension the same. Elements along the last dimension contain the
unions.