有人能帮我解释一下ano教程中的一行代码吗?

2024-07-05 12:30:29 发布

您现在位置:Python中文网/ 问答频道 /正文

在ano教程中提供的logistic regression example中,negative_log_likelihood函数中有一行代码,如下所示:

def negative_log_likelihood(self, y):
    """Return the mean of the negative log-likelihood of the prediction
    of this model under a given target distribution.

    .. math::

        \frac{1}{|\mathcal{D}|} \mathcal{L} (\theta=\{W,b\}, \mathcal{D}) =
        \frac{1}{|\mathcal{D}|} \sum_{i=0}^{|\mathcal{D}|} \log(P(Y=y^{(i)}|x^{(i)}, W,b)) \\
            \ell (\theta=\{W,b\}, \mathcal{D})

    :type y: theano.tensor.TensorType
    :param y: corresponds to a vector that gives for each example the
              correct label

    Note: we use the mean instead of the sum so that
          the learning rate is less dependent on the batch size
    """
    # y.shape[0] is (symbolically) the number of rows in y, i.e.,
    # number of examples (call it n) in the minibatch
    # T.arange(y.shape[0]) is a symbolic vector which will contain
    # [0,1,2,... n-1] T.log(self.p_y_given_x) is a matrix of
    # Log-Probabilities (call it LP) with one row per example and
    # one column per class LP[T.arange(y.shape[0]),y] is a vector
    # v containing [LP[0,y[0]], LP[1,y[1]], LP[2,y[2]], ...,
    # LP[n-1,y[n-1]]] and T.mean(LP[T.arange(y.shape[0]),y]) is
    # the mean (across minibatch examples) of the elements in v,
    # i.e., the mean log-likelihood across the minibatch.
    return -T.mean(T.log(self.p_y_given_x)[T.arange(y.shape[0]), y])

有人能帮我解释一下上面代码最后一行中方括号的用法吗?如何解释[T.arange(y.shape[0]), y]?在

谢谢!在


Tags: oftheselflogisexamplemeangiven
2条回答

我也对矩阵切片感到困惑。T、 arange(y.shape[0])是一个1d列表。y、 形状[0]取决于您设置的小批量的大小。y是与T.arange(y.shape[0])具有相同维度的标签列表。因此,根据@William Denman的参考,这个切片意味着:对于T.log(self.p_y_give_x)矩阵中的每一行,我们选择一个列索引y(其中y表示黄金标签,这里也使用它作为索引)。在

函数的注释中包含了所需的大部分信息。在

T.log(self.p_y_give_x)返回numpy矩阵。在

所以[T.arange(y.shape[0]),y]是矩阵的一个切片。这里我们使用的是numpy高级切片。参见:http://docs.scipy.org/doc/numpy/reference/arrays.indexing.html

相关问题 更多 >