Hello. I am doing a semantic segmentation task.I have super high resolution image and its corresponding mask.I downsampled the images so that the data could fit in the gpu memory. My network takes this downsampled image and output a score map. But there's more. My network keeps on upsampling to output a score map of the original super-high resolution. So how do I apply loss both between the intermediate score map and the down-sampled mask and the final score map and the orignal mask.(This is a technique called deep supervision)?
I read the document. There is "forward(dlNetwork)" function available. But that only supports one loss. I want the two loss combined together.