Added an "add_to" option to tt:copy_tensor(). There was also a bug in the
concat layer's backward() method. It was assigning the gradient to previous layers instead of adding the gradient, as required by the layer interface specification. This change also noticeably speeds up concat layers since only one CUDA kernel launch now happens per concat operation, rather than one kernel launch for each sample in a tensor.
Showing
Please
register
or
sign in
to comment