-
Notifications
You must be signed in to change notification settings - Fork 0
/
papers.json
1682 lines (1682 loc) · 181 KB
/
papers.json
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
{
"1": {
"abstract": "We propose a novel weakly supervised learning segmentation based on several global constraints derived from box annotations. Particularly, we leverage a classical tightness prior to a deep learning setting via imposing a set of constraints on the network outputs. Such a powerful topological prior prevents solutions from excessive shrinking by enforcing any horizontal or vertical line within the bounding box to contain, at least, one pixel of the foreground region. Furthermore, we integrate our deep tightness prior with a global background emptiness constraint, guiding training with information outside the bounding box. We demonstrate experimentally that such a global constraint is much more powerful than standard cross-entropy for the background class. Our optimization problem is challenging as it takes the form of a large set of inequality constraints on the outputs of deep networks. We solve it with sequence of unconstrained losses based on a recent powerful extension of the log-barrier method, which is well-known in the context of interior-point methods. This accommodates standard stochastic gradient descent (SGD) for training deep networks, while avoiding computationally expensive and unstable Lagrangian dual steps and projections. Extensive experiments over two different public data sets and applications (prostate and brain lesions) demonstrate that the synergy between our global tightness and emptiness priors yield very competitive performances, approaching full supervision and outperforming significantly DeepCut. Furthermore, our approach removes the need for computationally expensive proposal generation. Our code is shared anonymously. ",
"authors": "Hoel Kervadec, Jose Dolz, Shanshan Wang, Eric Granger, Ismail Ben Ayed",
"award": "",
"id": "1",
"or_id": "m7HZ-yil_-",
"oral": "True",
"pmlr_url": "http://proceedings.mlr.press/v121/kervadec20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/kervadec20.pdf",
"title": "Bounding boxes for weakly supervised segmentation: Global constraints get close to full supervision",
"url": "papers/kervadec20.html",
"yt_full": "1HIK_z-XeMU",
"yt_teaser": ""
},
"2": {
"abstract": "Incorporating distance transform maps of ground truth into segmentation CNNs has been an interesting new trend in the last year. Despite many great works leading to improvements on a variety of segmentation tasks, the comparison among these methods has not been well studied.\n In this paper, our \\emph{first contribution} is to summarize the latest developments of these methods in the 3D medical segmentation field.\n The \\emph{second contribution} is that we systematically evaluated five benchmark methods on two representative public datasets.\n These experiments highlight that all the five benchmark methods can bring performance gains to baseline V-Net. However, the implementation details have a noticeable impact on the performance, and not all the methods hold the benefits in different datasets.\n Finally, we suggest the best practices and indicate unsolved problems for incorporating distance transform maps into CNNs, which we hope will be useful for the community. The codes and trained models are publicly available at \\url{https://github.com/JunMa11/SegWithDistMap}.",
"authors": "Jun Ma, Zhan Wei, Yiwen Zhang, Yixin Wang, Rongfei Lv, Cheng Zhu, Gaoxiang Chen, Jianan Liu, Chao Peng, Lei Wang, Yunpeng Wang, Jianan Chen",
"award": "",
"id": "2",
"or_id": "hM4pNbXWst",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/ma20b.html",
"schedule": "",
"short": "False",
"slides": "/slides/ma20a.pdf",
"title": "How Distance Transform Maps Boost Segmentation CNNs: An Empirical Study",
"url": "papers/ma20a.html",
"yt_full": "lqfGw0tDy3k",
"yt_teaser": "G23FofuAOFM"
},
"6": {
"abstract": "A learning-based posterior distribution estimation method, Probabilistic Dipole Inversion (PDI), is proposed to solve quantitative susceptibility mapping (QSM) inverse problem in MRI with uncertainty estimation. A deep convolutional neural network (CNN) is used to represent the multivariate Gaussian distribution as the approximated posterior distribution of susceptibility given the input measured field. In PDI, such CNN is firstly trained on healthy subjects' data with labels by maximizing the posterior Gaussian distribution loss function as used in Bayesian deep learning. When testing on each patient' data without any label, PDI updates the pre-trained CNN's weights in an unsupervised fashion by minimizing the Kullback\u2013Leibler divergence between the approximated posterior distribution represented by CNN and the true posterior distribution given the likelihood distribution from known physical model and pre-defined prior distribution. Based on our experiments, PDI provides additional uncertainty estimation compared to the conventional MAP approach, meanwhile addressing the potential discrepancy issue of CNN when test data deviates from training dataset.",
"authors": "Jinwei Zhang, Hang Zhang, Mert Sabuncu, Pascal Spincemaille, Thanh Nguyen, Yi Wang",
"award": "",
"id": "6",
"or_id": "DuWrLOZ27k",
"oral": "True",
"pmlr_url": "http://proceedings.mlr.press/v121/zhang20b.html",
"schedule": "",
"short": "False",
"slides": "/slides/zhang20a.pdf",
"title": "Bayesian Learning of Probabilistic Dipole Inversion for Quantitative Susceptibility Mapping",
"url": "papers/zhang20a.html",
"yt_full": "v9f6DjIbDbI",
"yt_teaser": ""
},
"9": {
"abstract": "Deep learning networks are being developed in every stage of the MRI workflow and have provided state-of-the-art results. However, this has come at the cost of increased computation requirement and storage. Hence, replacing the networks with compact models at various stages in the MRI workflow can significantly reduce the required storage space and provide considerable speedup. In computer vision, knowledge distillation is a commonly used method for model compression. In our work, we propose a knowledge distillation (KD) framework for the image to image problems in the MRI workflow in order to develop compact, low-parameter models without a significant drop in performance. We propose a combination of the attention-based feature distillation method and imitation loss and demonstrate its effectiveness on the popular MRI reconstruction architecture, DC-CNN. We conduct extensive experiments using Cardiac, Brain, and Knee MRI datasets for 4x, 5x and 8x accelerations. We observed that the student network trained with the assistance of the teacher using our proposed KD framework provided significant improvement over the student network trained without assistance across all the datasets and acceleration factors. Specifically, for the Knee dataset, the student network achieves $65\\%$ parameter reduction, 2x faster CPU running time, and 1.5x faster GPU running time compared to the teacher. Furthermore, we compare our attention-based feature distillation method with other feature distillation methods. We also conduct an ablative study to understand the significance of attention-based distillation and imitation loss. We also extend our KD framework for MRI super-resolution and show encouraging results. ",
"authors": "Balamurali Murugesan, Sricharan Vijayarangan, Kaushik Sarveswaran, Keerthi Ram, Mohanasankar Sivaprakasam",
"award": "",
"id": "9",
"or_id": "OrBdiT86_O",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/murugesan20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/murugesan20.pdf",
"title": "KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow",
"url": "papers/murugesan20.html",
"yt_full": "vYiBl9hiNGw",
"yt_teaser": "ue3XdYiCDz8"
},
"11": {
"abstract": "Differentiable programming is able to combine different functions or modules in a data processing pipeline with the goal of applying gradient descent-based end-to-end learning or optimization. A significant impediment to differentiable programming is the non-differentiable nature of some functions. We propose to overcome this difficulty by using neural networks to approximate such modules. An approximating neural network provides synthetic gradients (SG) for backpropagation across a non-differentiable module. Our design is grounded on a well-known theory that gradient of an approximating neural network can approximate a sub-gradient of a weakly differentiable function. We apply SG to combine convolutional neural network (CNN) with dynamic programming (DP) in end-to-end learning for segmenting left ventricle from short axis view of heart MRI. Our experiments show that end-to-end combination of CNN and DP requires fewer labeled images to achieve a significantly better segmentation accuracy than using only CNN.",
"authors": "Nhat M. Nguyen, Nilanjan Ray",
"award": "",
"id": "11",
"or_id": "_4_RPMYWN",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/nguyen20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/nguyen20a.pdf",
"title": "End-to-end learning of convolutional neural net and dynamic programming for left ventricle segmentation",
"url": "papers/nguyen20a.html",
"yt_full": "16gEbSz65YM",
"yt_teaser": "X0hBhXi-laY"
},
"13": {
"abstract": "Optical coherence tomography (OCT) is a non-invasive imaging technology that can provide micrometer-resolution cross-sectional images of the inner structures of the eye. It is widely used for the diagnosis of ophthalmic diseases with retinal alteration, such as layer deformation and fluid accumulation. In this paper, a novel framework was proposed to segment retinal layers with fluid presence. The main contribution of this study is two folds: 1) we developed a cascaded network framework to incorporate the prior structural knowledge; 2) we proposed a novel deep neural network based on U-Net and fully convolutional network, termed LF-UNet. Cross validation experiments proved that the proposed LF-UNet has superior performance comparing with the state-of-the-art methods, and incorporating the relative distance map structural prior information could further improve the performance regardless of the network.",
"authors": "Da Ma, Donghuan Lu, Morgan Heisler, Setareh Dabiri, Sieun Lee, Gavin Weiguan Ding, Marinko V. Sarunic, Mirza Faisal Beg",
"award": "",
"id": "13",
"or_id": "dxVMXBzKKQ",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/ma20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/ma20b.pdf",
"title": "Cascade Dual-branch Deep Neural Networks for Retinal Layer and fluid Segmentation of Optical Coherence Tomography Incorporating Relative Positional Map",
"url": "papers/ma20b.html",
"yt_full": "jFMWE_-nZv0",
"yt_teaser": "fRb6s84YSew"
},
"14": {
"abstract": "Fully-automatic lung lobe segmentation is challenging due to anatomical variations, pathologies, and incomplete fissures. We trained a 3D u-net for pulmonary lobe segmentation on 49 mainly publically available datasets and introduced a weighted Dice loss function to emphasize the lobar boundaries. To validate the performance of the proposed method we compared the results to two other methods. The new loss function improved the mean distance to 1.46 mm (compared to 2.08 mm for simple loss function without weighting).",
"authors": "Bianca Lassen-Schmidt, Alessa Hering, Stefan Krass, Hans Meine",
"award": "",
"id": "14",
"or_id": "AkziGgmwl",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/lassen-schmidt20.pdf",
"title": "Automatic segmentation of the pulmonary lobes with a 3D u-net and optimized loss function",
"url": "papers/lassen-schmidt20.html",
"yt_full": "_OMrnUQcUUs",
"yt_teaser": "IfeHzhAK1yM"
},
"17": {
"abstract": "Segmentation of cell nuclei in microscopy images is a prevalent necessity in cell biology.\n Especially for three-dimensional datasets, manual segmentation is prohibitively time-consuming, motivating the need for automated methods. Learning-based methods trained on pixel-wise ground-truth segmentations have been shown to yield state-of-the-art results on 2d benchmark image data of nuclei, yet a respective benchmark is missing for 3d image data. In this work, we perform a comparative evaluation of nuclei segmentation algorithms on a database of manually segmented 3d light microscopy volumes. We propose a novel learning strategy that boosts segmentation accuracy by means of a simple auxiliary task, thereby robustly outperforming each of our baselines. Furthermore, we show that one of our baselines, the popular three-label model, when trained with our proposed auxiliary task, outperforms the recent StarDist-3D.\n As an additional, practical contribution, we benchmark nuclei segmentation against nuclei detection, i.e. the task of merely pinpointing individual nuclei without generating respective pixel-accurate segmentations. For learning nuclei detection, large 3d training datasets of manually annotated nuclei center points are available. However, the impact on detection accuracy caused by training on such sparse ground truth as opposed to dense pixel-wise ground truth has not yet been quantified. To this end, we compare nuclei detection accuracy yielded by training on dense vs. sparse ground truth. Our results suggest that training on sparse ground truth yields competitive nuclei detection rates. ",
"authors": "Peter Hirsch, Dagmar Kainmueller",
"award": "",
"id": "17",
"or_id": "iJVionbWNX",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/hirsch20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/hirsch20.pdf",
"title": "An Auxiliary Task for Learning Nuclei Segmentation in 3D Microscopy Images",
"url": "papers/hirsch20.html",
"yt_full": "9HfD3vEXtHE",
"yt_teaser": "fLaZoYgUpOo"
},
"18": {
"abstract": "Undersampled MR image recovery has been widely studied for accelerated MR acquisition. However, it has been mostly studied under a single sequence scenario, despite the fact that multi-sequence MR scan is common in practice. In this paper, we aim to optimize multi-sequence MR image recovery from undersampled k-space data under an overall time constraint while considering the difference in acquisition time for various sequences. We first formulate it as a constrained optimization problem and then show that finding the optimal sampling strategy for all sequences and the best recovery model at the same time is combinatorial and hence computationally prohibitive. To solve this problem, we propose a blind recovery model that simultaneously recovers multiple sequences, and an efficient approach to find proper combination of sampling strategy and recovery model. Our experiments demonstrate that the proposed method outperforms sequence-wise recovery, and sheds light on how to decide the undersampling strategy for sequences within an overall time budget.",
"authors": "Cheng Peng, Wei-An Lin, Rama Chellappa, S. Kevin Zhou",
"award": "",
"id": "18",
"or_id": "Pk7In-gVEd",
"oral": "True",
"pmlr_url": "http://proceedings.mlr.press/v121/peng20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/peng20b.pdf",
"title": "Towards multi-sequence MR image recovery from undersampled k-space data",
"url": "papers/peng20b.html",
"yt_full": "nK-JLpor-vM",
"yt_teaser": ""
},
"23": {
"abstract": "We present a new approach based on deep convolutional neural networks (CNNs) for predicting the presence of 14 common thoracic diseases and observations. A strong set of CNNs are trained on over 200,000 chest X-ray images provided by CheXpert - a large scale chest X-ray dataset. In particular, dependencies among abnormality labels and uncertain samples are fully exploited during the training and inference stages. Experiments indicate that the proposed method achieves a mean area under the curve (AUC) of 0.940 in predicting 5 selected pathologies. To the best of our knowledge, this is the highest AUC score yet reported on this dataset to date. Additionally, the proposed method is also evaluated on the independent test set of the CheXpert competition and reports a performance level comparable to practicing radiologists. Our obtained result ranks first on the CheXpert leaderboard at the time of writing this paper.",
"authors": "Hieu H. Pham, Tung T. Le, Dat T. Ngo, Dat Q. Tran, Ha Q. Nguyen",
"award": "",
"id": "23",
"or_id": "4o1GLIIHlh",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/pham20.pdf",
"title": "Interpreting Chest X-rays via CNNs that Exploit Hierarchical Disease Dependencies and Uncertainty Labels",
"url": "papers/pham20.html",
"yt_full": "gR8kvihcSk0",
"yt_teaser": "oOBRZmyJW7o"
},
"27": {
"abstract": "There is an urgent need for a paradigm shift from group-wise comparisons to individual diagnosis in diffusion MRI (dMRI) to enable the analysis of rare cases and clinically-heterogeneous groups. Deep autoencoders have shown great potential to detect anomalies in neuroimaging data. We present a framework that operates on the manifold of white matter (WM) pathways to learn normative microstructural features, and discriminate those at genetic risk from controls in a paediatric population. ",
"authors": "Maxime Chamberland, Sila Genc, Erika P. Raven, Greg D. Parker, Adam Cunningham, Joanne Doherty, Marianne van den Bree, Chantal M. W. Tax, Derek K. Jones",
"award": "",
"id": "27",
"or_id": "heX-Rk0TE0",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/chamberland20.pdf",
"title": "Tractometry-based Anomaly Detection for Single-subject White Matter Analysis",
"url": "papers/chamberland20.html",
"yt_full": "ydTnrTbaQYA",
"yt_teaser": "j9m9pqqQZHk"
},
"28": {
"abstract": "Deep learning models have been shown to have success in reconstructing accelerated MRI, over traditional methods. However, it has been observed that these methods tend to miss the small features that are rare, such as meniscal tears, subchondral osteophyte, etc. This is a concerning finding as these small and rare features are the most relevant in clinical diagnostic settings. In this work, we propose a framework to find the worst-case false negatives by adversarially attacking the trained models and improve the models' ability to reconstruct the small features by robust training.",
"authors": "Kaiyang Cheng, Francesco Caliv\u00e1, Rutwik Shah, Misung Han, Sharmila Majumdar, Valentina Pedoia",
"award": "Best paper Award",
"id": "28",
"or_id": "7NF2rZwE-z",
"oral": "True",
"pmlr_url": "http://proceedings.mlr.press/v121/cheng20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/cheng20.pdf",
"title": "Addressing The False Negative Problem of Deep Learning MRI Reconstruction Models by Adversarial Attacks and Robust Training",
"url": "papers/cheng20.html",
"yt_full": "7ccE6SXy9t8",
"yt_teaser": ""
},
"33": {
"abstract": "Signed distance map (SDM) is a common representation of surfaces in medical image analysis and machine learning. The computational complexity of SDM for 3D parametric shapes is often a bottleneck in many applications, thus limiting their interest. In this paper, we propose a learning-based SDM generation neural network which is demonstrated on a tridimensional cochlea shape model parameterized by 4 shape parameters.\n\n The proposed SDM Neural Network generates a cochlea signed distance map depending on four input parameters and we show that the deep learning approach leads to a 60 fold improvement in the time of computation compared to more classical SDM generation methods. Therefore, the proposed approach achieves a good trade-off between accuracy and efficiency. ",
"authors": "Zihao Wang, Clair Vandersteen, Thomas Demarcy, Dan Gnansia, Charles Raffaelli, Nicolas Guevara, Herve\u0301 Delingette",
"award": "",
"id": "33",
"or_id": "b2N5ZuEouu",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/wang20a.pdf",
"title": "A Deep Learning based Fast Signed Distance Map Generation",
"url": "papers/wang20a.html",
"yt_full": "_o8KzH0ll3o",
"yt_teaser": "BeWMzeZIjEI"
},
"35": {
"abstract": "Most deep learning models in chest X-ray prediction utilize the posteroanterior (PA) view due to the lack of other views available. PadChest is a large-scale chest X-ray dataset that has almost 200 labels and multiple views available. In this work, we use PadChest to explore multiple approaches to merging the PA and lateral views for predicting the radiological labels associated with the X-ray image. We find that different methods of merging the model utilize the lateral view differently. We also find that including the lateral view increases performance for 32 labels in the dataset, while being neutral for the others.\n The increase in overall performance is comparable to the one obtained by using only the PA view with twice the amount of patients in the training set.",
"authors": "Mohammad Hashir, Hadrien Bertrand, Joseph Paul Cohen",
"award": "",
"id": "35",
"or_id": "rY3bgRRHnD",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/hashir20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/hashir20.pdf",
"title": "Quantifying the Value of Lateral Views in Deep Learning for Chest X-rays",
"url": "papers/hashir20.html",
"yt_full": "5VjZyIR1224",
"yt_teaser": "___k6jBuPEo"
},
"36": {
"abstract": "This large scale study focuses on quantifying what X-rays diagnostic prediction tasks generalize well across multiple different datasets. We present evidence that the issue of generalization is not due to a shift in the images but instead a shift in the labels. We study the cross-domain performance, agreement between models, and model representations. We find interesting discrepancies between performance and agreement where models which both achieve good performance disagree in their predictions as well as models which agree yet achieve poor performance. We also test for concept similarity by regularizing a network to group tasks across multiple datasets together and observe variation across the tasks.",
"authors": "Joseph Paul Cohen, Mohammad Hashir, Rupert Brooks, Hadrien Bertrand",
"award": "",
"id": "36",
"or_id": "RZ-1WCgOQU",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/cohen20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/cohen20.pdf",
"title": "On the limits of cross-domain generalization in automated X-ray prediction",
"url": "papers/cohen20.html",
"yt_full": "TOPu7AWZSG8",
"yt_teaser": "LZBFwkAB7f8"
},
"37": {
"abstract": "Over the last years, Deep Learning has been successfully applied to a broad range of medical applications. Especially in the context of chest X-ray classification, results have been reported which are on par, or even superior to experienced radiologists. Despite this success in controlled experimental environments, it has been noted that the ability of Deep Learning models to generalize to data from a new domain (with potentially different tasks) is often limited. In order to address this challenge, we investigate techniques from the field of Continual Learning (CL) including Joint Training (JT), Elastic Weight Consolidation (EWC) and Learning Without Forgetting (LWF). Using the ChestX-ray14 and the MIMIC-CXR datasets, we demonstrate empirically that these methods provide promising options to improve the performance of Deep Learning models on a target domain and to mitigate effectively catastrophic forgetting for the source domain. To this end, the best overall performance was obtained using JT, while for LWF competitive results could be achieved - even without accessing data from the source domain.",
"authors": "Matthias Lenga, Heinrich Schulz, Axel Saalbach",
"award": "",
"id": "37",
"or_id": "lgE-MUlU1g",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/lenga20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/lenga20.pdf",
"title": "Continual Learning for Domain Adaptation in Chest X-ray Classification",
"url": "papers/lenga20.html",
"yt_full": "qzN4zNk8-Tc",
"yt_teaser": "bjuglG8y3q0"
},
"38": {
"abstract": "For the majority of the learning-based segmentation methods, a large quantity of high-quality training data is required. In this paper, we present a novel learning-based segmentation model that could be trained semi- or un- supervised. Specifically, in the unsupervised setting, we parameterize the Active contour without edges (ACWE) framework via a convolutional neural network (ConvNet), and optimize the parameters of the ConvNet using a self-supervised method. In another setting (semi-supervised), the auxiliary segmentation ground truth is used during training. We show that the method provides fast and high-quality bone segmentation in the context of single-photon emission computed tomography (SPECT) image.",
"authors": "Junyu Chen, Eric C. Frey",
"award": "",
"id": "38",
"or_id": "XrbnSCv4LU",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/chen20.pdf",
"title": "Medical Image Segmentation via Unsupervised Convolutional Neural Network",
"url": "papers/chen20.html",
"yt_full": "IVMy-EUeTc8",
"yt_teaser": "JOa6zS7TNCI"
},
"41": {
"abstract": "We propose a novel method that combines a convolutional neural network (CNN) with a long short-term memory (LSTM) mechanism for accurate prediction of intracranial hemorrhage on computed tomography (CT) scans. The CNN plays the role of a slice-wise feature extractor while the LSTM is responsible for linking the features across slices. The whole architecture is trained end-to-end with input being an RGB-like image formed by stacking 3 different viewing windows of a single slice. We validate the method on the recent RSNA Intracranial Hemorrhage Detection challenge and on the CQ500 dataset. For the RSNA challenge, our best single model achieves a weighted log loss of 0.0529 on the leaderboard, which is comparable to the top 3\\% performances, almost all of which make use of ensemble learning. Importantly, our method generalizes very well: the model trained on the RSNA dataset significantly outperforms the 2D model, which does not take into account the relationship between slices, on CQ500. Our codes and models will be made public.",
"authors": "Nhan T. Nguyen, Dat Q. Tran, Nghia T. Nguyen, Ha Q. Nguyen",
"award": "",
"id": "41",
"or_id": "1IoPbyuPFT",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/nguyen20b.pdf",
"title": "A CNN-LSTM Architecture for Detection of Intracranial Hemorrhage on CT scans",
"url": "papers/nguyen20b.html",
"yt_full": "BI2PBdDFK10",
"yt_teaser": "rZuYAKt26fo"
},
"42": {
"abstract": "Pulmonary Embolism (PE) is a life-threatening disorder associated with high mortality and morbidity. Prompt diagnosis and immediate initiation of therapeutic action is important. We explored a deep learning model to detect PE on volumetric contrast-enhanced chest CT scans using a 2-stage training strategy. First, a residual convolutional neural network (ResNet) was trained using annotated 2D images. In addition to the classification loss, an attention loss was added during training to help the network focus attention on PE. Next, a recurrent network was used to scan sequentially through the features provided by the pre-trained ResNet to detect PE. This combination allows the network to be trained using both a limited and sparse set of pixel-level annotated images and a large number of easily obtainable patient-level image-label pairs. We used 1,670 sparsely annotated studies and more than 10,000 labeled studies in our training. On a test set with 2,160 patient studies, the proposed method achieved an area under the ROC curve (AUC) of 0.812. The proposed framework is also able to provide localized attention maps that indicate possible PE lesions, which could potentially help radiologists accelerate the diagnostic process.",
"authors": "Luyao Shi, Deepta Rajan, Shafiq Abedin, David Beymer, Ehsan Dehghan",
"award": "",
"id": "42",
"or_id": "hsGCHJDRm2",
"oral": "True",
"pmlr_url": "http://proceedings.mlr.press/v121/shi20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/shi20.pdf",
"title": "Automatic Diagnosis of Pulmonary Embolism Using an Attention-guided Framework: A Large-scale Study",
"url": "papers/shi20.html",
"yt_full": "gg9f4XEScec",
"yt_teaser": "GSQIJjNgytQ"
},
"43": {
"abstract": "We consider the problem of domain generalization, namely, how to learn representations given data from a set of domains that generalize to data from a previously unseen domain. We propose the Domain Invariant Variational Autoencoder (DIVA), a generative model that tackles this problem by learning three independent latent subspaces, one for the domain, one for the class, and one for any residual variations. We highlight that due to the generative nature of our model we can also incorporate unlabeled data from known or previously unseen domains. To the best of our knowledge this has not been done before in a domain generalization setting. This property is highly desirable in fields like medical imaging where labeled data is scarce. We experimentally evaluate our model on the rotated MNIST benchmark and a malaria cell images dataset where we show that (i) the learned subspaces are indeed complementary to each other, (ii) we improve upon recent works on this task and (iii) incorporating unlabelled data can boost the performance even further.",
"authors": "Maximilian Ilse, Jakub M. Tomczak, Christos Louizos, Max Welling",
"award": "",
"id": "43",
"or_id": "RmNckVums7",
"oral": "True",
"pmlr_url": "http://proceedings.mlr.press/v121/ilse20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/ilse20.pdf",
"title": "DIVA: Domain Invariant Variational Autoencoders",
"url": "papers/ilse20.html",
"yt_full": "1GBSjGtr2rE",
"yt_teaser": "pLSElJxW7vM"
},
"45": {
"abstract": "Minimization of distribution matching losses is a principled approach to domain adaptation in the context of image classification. However, it is largely overlooked in adapting segmentation networks, which is currently dominated by adversarial models. We propose a class of loss functions, which encourage direct kernel density matching in the network-output space, up to some geometric transformations computed from unlabeled inputs. Rather than using an intermediate domain discriminator, our direct approach unifies distribution matching and segmentation in a single loss. Therefore, it simplifies segmentation adaptation by avoiding extra adversarial steps, while improving quality, stability and efficiency of training. We juxtapose our approach to state-of-the-art segmentation adaptation via adversarial training in the network-output space. In the challenging task of adapting brain segmentation across different magnetic resonance imaging (MRI) modalities, our approach achieves significantly better results both in terms of accuracy and stability.\n ",
"authors": "Georg Pichler, Jose Dolz, Ismail Ben Ayed, Pablo Piantanida",
"award": "",
"id": "45",
"or_id": "-C9f-eGAuV",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/pichler20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/pichler20.pdf",
"title": "On Direct Distribution Matching for Adapting Segmentation Networks",
"url": "papers/pichler20.html",
"yt_full": "OX1qKZ8NR8I",
"yt_teaser": "6RQxDpWY6OE"
},
"47": {
"abstract": "Spiral acquisitions are preferred in real-time MRI because of their time efficiency. A fundamental limitation of spirals is image blurring due to off-resonance, which degrades image quality significantly at air-tissue boundaries. Here, we demonstrate a simple CNN-based deblurring method for spiral real-time MRI of human speech production. We show the CNN-based deblurring is capable of restoring blurred vocal tract tissue boundaries, without a need for exam-specific field maps. Deblurring performance is superior to a current auto-calibrated method, and slightly inferior to ideal reconstruction with perfect knowledge of the field maps. ",
"authors": "Yongwan Lim, Shrikanth Narayanan, Krishna Nayak",
"award": "",
"id": "47",
"or_id": "zYareJYs8Z",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/lim20.pdf",
"title": "Deblurring for spiral real-time MRI using convolutional neural networks",
"url": "papers/lim20.html",
"yt_full": "OtuZmOuJQlU",
"yt_teaser": "yNAMOvv9CyI"
},
"48": {
"abstract": "Deep learning has proven to be an essential tool for medical image analysis. However, the need for accurately labeled input data, often requiring time- and labor-intensive annotation by experts, is a major limitation to the use of deep learning. One solution to this challenge is to allow for use of coarse or noisy labels, which could permit more efficient and scalable labeling of images. In this work, we develop a lopsided loss function based on entropy regularization that assumes the existence of a nontrivial false negative rate in the target annotations. Starting with a carefully annotated brain metastasis lesion dataset, we simulate data with false negatives by (1) randomly censoring the annotated lesions and (2) systematically censoring the smallest lesions. The latter better models true physician error because smaller lesions are harder to notice than the larger ones. Even with a simulated false negative rate as high as 50%, applying our loss function to randomly censored data preserves maximum sensitivity at 97% of the baseline with uncensored training data, compared to just 10% for a standard loss function. For the size-based censorship, performance is restored from 17% with the current standard to 88% with our lopsided bootstrap loss. Our work will enable more efficient scaling of the image labeling process, in parallel with other approaches on creating more efficient user interfaces and tools for annotation.",
"authors": "Darvin Yi, Endre Gr\u00f8vik, Michael Iv, Elizabeth Tong, Greg Zaharchuk, Daniel Rubin",
"award": "",
"id": "48",
"or_id": "55VoQFvQM",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/yi20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/yi20.pdf",
"title": "Brain Metastasis Segmentation Network Trained with Robustness to Annotations with Multiple False Negatives",
"url": "papers/yi20.html",
"yt_full": "79xqFZ0aRN8",
"yt_teaser": "i88RD3VDwTs"
},
"53": {
"abstract": "Preliminary diagnosis of fungal infections can rely on microscopic examination. However, in many cases, it does not allow unambiguous identification of the species due to their visual similarity. Therefore, it is usually necessary to use additional biochemical tests. That involves additional costs and extends the identification process up to 10 days. Such a delay in the implementation of targeted therapy may be grave in consequence as the mortality rate for immunosuppressed patients is high. In this paper, we apply a machine learning approach based on deep neural networks and bag-of-words to classify microscopic images of various fungi species. Our approach makes the last stage of biochemical identification redundant, shortening the identification process by 2-3 days, and reducing the cost of the diagnosis.",
"authors": "Bartosz Zieli\u0144ski, Agnieszka Sroka-Oleksiak, Dawid Rymarczyk, Adam Piekarczyk, Monika Brzychczy-W\u0142och",
"award": "",
"id": "53",
"or_id": "AEhp_Cqq-h",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/zielinski20.pdf",
"title": "Deep learning approach to describing and classifying fungi microscopic images",
"url": "papers/zielinski20.html",
"yt_full": "5UuhAat2qig",
"yt_teaser": "xCFpZI50cTg"
},
"54": {
"abstract": "Accurate segmentation of spine Magnetic Resonance Imaging (MRI) is highly demanded in morphological research, quantitative analysis, and diseases identification, such as spinal canal stenosis, disc herniation and degeneration. However, accurate spine segmentation is challenging because of the irregular shape, artifacts and large variability between slices. To alleviate these problems, spatial information is used for more continuous and accurate segmentation such as by 3D convolutional neural networks (CNN) . However, 3D CNN suffers from higher computational cost, memory cost and risk of over-fitting, especially for medical images where the number of labeled data is limited. To address these problems, we apply the attention mechanism for the utilization of inter-slice information in 3D segmentation tasks based on 2D convolutional networks and propose a spatial attention-based densely connected U-Net (SAU-Net), which consists of Dense U-Net for extraction of intra-slice features and an inter-slice attention module (ISA) to utilize inter-slice information from adjacent slices and refine the segmentation results. Experimental results demonstrate the effectiveness of ISA as well as higher accuracy and efficiency of segmentation results of our method compared with other deep learning methods.",
"authors": "Yichi Zhang, Lin Yuan, Yujia Wang, Jicong Zhang",
"award": "",
"id": "54",
"or_id": "wlszIiXbfS",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/zhang20c.html",
"schedule": "",
"short": "False",
"slides": "/slides/zhang20b.pdf",
"title": "SAU-Net: Efficient 3D Spine MRI Segmentation Using Inter-Slice Attention",
"url": "papers/zhang20b.html",
"yt_full": "tzlzdE_ANuc",
"yt_teaser": "vqV45k9Wjxc"
},
"56": {
"abstract": "Convolutional Neural network based MR reconstruction methods have shown to provide fast and high quality reconstructions. A primary drawback with a CNN-based model is that it lacks flexibility and can effectively operate only for a specific acquisition context limiting practical applicability. By acquisition context, we mean a specific combination of three input settings considered namely, the anatomy under study, undersampling mask pattern and acceleration factor for undersampling. The model could be trained jointly on images combining multiple contexts. However the model does not meet the performance of context specific models nor extensible to contexts unseen at train time. This necessitates a modification to the existing architecture in generating context specific weights so as to incorporate flexibility to multiple contexts. We propose a multiple acquisition context based network, called MAC-ReconNet for MRI reconstruction, flexible to multiple acquisition contexts and generalizable to unseen contexts for applicability in real scenarios. The proposed network has an MRI reconstruction module and a dynamic weight prediction (DWP) module. The DWP module takes the corresponding acquisition context information as input and learns the context-specific weights of the reconstruction module which changes dynamically with context at run time. We show that the proposed approach can handle multiple contexts based on Cardiac and Brain datasets, Gaussian and Cartesian undersampling patterns and five acceleration factors. The proposed network outperforms the naive jointly trained model and gives competitive results with the context-specific models both quantitatively and qualitatively. We also demonstrate the generalizability of our model by testing on contexts unseen at train time.",
"authors": "Sriprabha Ramanarayanan, Balamurali Murugesan, Keerthi Ram, Mohanasankar Sivaprakasam",
"award": "Runner-up for best paper Award",
"id": "56",
"or_id": "pMHk_HIZf7",
"oral": "True",
"pmlr_url": "http://proceedings.mlr.press/v121/ramanarayanan20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/ramanarayanan20.pdf",
"title": "MAC-ReconNet: A Multiple Acquisition Context based Convolutional Neural Network for MR Image Reconstruction using Dynamic Weight Prediction",
"url": "papers/ramanarayanan20.html",
"yt_full": "HQ1ioGUwhcE",
"yt_teaser": "_rzsMOcyV3o"
},
"57": {
"abstract": "In this work we propose a multi-scale recurrent encoder-decoder architecture to predict the breathing induced organ deformation in future frames. The model was trained end-to-end from input images to predict a sequence of motion labels. Targets were created by quantizing the motion fields obtained from deformable image registration. We propose a multi-scale feature extraction scheme in the spatial encoder which processes the input at different resolutions. We report results using MRI free-breathing acquisitions from 12 volunteers. Experiments were aimed at investigating the proposed multi-scale design and the effect of increasing the number of predicted frames on the overall accuracy of the model. The proposed model was able to predict vessel positions in the next temporal image with a mean accuracy of 2.03 (2.89) mm showing increased performance in comparison with state-of-the-art approaches.",
"authors": "Liset Vazquez Romaguera, Rosalie Plantefeve, Samuel Kadoury",
"award": "",
"id": "57",
"or_id": "901HZmWDHH",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/romaguera20.pdf",
"title": "Spatiotemporal motion prediction in free-breathing liver scans via a recurrent multi-scale encoder decoder",
"url": "papers/romaguera20.html",
"yt_full": "Yf8q0Nq9QQc",
"yt_teaser": "zjI5M5zf8do"
},
"58": {
"abstract": "Radiogenomic map linking image features and gene expression profiles has great potential for non-invasively identifying molecular properties of a particular type of disease. Conventionally, such map is produced in three independent steps: 1) gene-clustering to metagenes, 2) image feature extraction, and 3) statistical correlation between metagenes and image features. Each step is separately performed and relies on arbitrary measurements without considering the correlation among each other. In this work, we investigate the potential of an end-to-end method fusing gene code with image features to generate synthetic pathology image and learn radiogenomic map simultaneously. To achieve this goal, we develop a multi-conditional generative adversarial network (GAN) conditioned on both background images and gene expression code, synthesizing the corresponding image. Image and gene features are fused at different scales to ensure both the separation of pathology part and background, as well as the realism and quality of the synthesized image. We tested our method on non-small cell lung cancer (NSCLC) dataset. Results demonstrate that the proposed method produces realistic synthetic images, and provides a promising way to find gene-image relationship in a holistic end-to-end manner.",
"authors": "Ziyue Xu, Xiaosong Wang, Hoo-Chang Shin, Dong Yang, Holger Roth, Fausto Milletari, Ling Zhang, Daguang Xu",
"award": "",
"id": "58",
"or_id": "2wAX1X5X6n",
"oral": "True",
"pmlr_url": "http://proceedings.mlr.press/v121/xu20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/xu20.pdf",
"title": "Correlation via Synthesis: End-to-end Image Generation and Radiogenomic Learning Based on Generative Adversarial Network",
"url": "papers/xu20.html",
"yt_full": "nwR4DzyIQdY",
"yt_teaser": ""
},
"59": {
"abstract": "We propose a dual-domain cascade of U-nets (i.e. a ``W-net\") operating in both the spatial frequency and image domains to enhance low-dose CT (LDCT) images without the need for proprietary x-ray projection data. The central slice theorem motivated the use of the spatial frequency domain in place of the raw sinogram. Data were obtained from the AAPM Low-dose Grand Challenge. A combination of Fourier space (F) and/or image domain (I) U-nets and W-nets were trained with a multi-scale structural similarity and mean absolute error loss function to denoise filtered back projected (FBP) LDCT images while maintaining perceptual features important for diagnostic accuracy. Deep learning enhancements were superior to FBP LDCT images in quantitative and qualitative performance with the dual-domain W-nets outperforming single-domain U-net cascades. Our results suggest that spatial frequency learning in conjunction with image-domain processing can produce superior LDCT enhancement than image-domain-only networks. ",
"authors": "Kevin J. Chung, Roberto Souza, Richard Frayne, Ting-Yim Lee",
"award": "",
"id": "59",
"or_id": "rw5BswbvMB",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/chung20.pdf",
"title": "Low-dose CT Enhancement Network with a Perceptual Loss Function in the Spatial Frequency and Image Domains",
"url": "papers/chung20.html",
"yt_full": "85XkQeBjoKo",
"yt_teaser": "_IR9dzn3tJ0"
},
"64": {
"abstract": "In this work, we present a novel unsupervised image registration algorithm. It is differentiable end-to-end and can be used for both multi-modal and mono-modal registration. This is done using mutual information (MI) as a metric. The novelty here is that rather than using traditional ways of approximating MI which are often histogram based, we use a neural estimator called MINE and supplement it with matrix exponential for transformation matrix computation. The introduction of MINE tackles some of the drawbacks of histogram based MI computation and matrix exponential makes the optimization process smoother. We also introduce the idea of a multi-resolution loss, which makes the optimization process faster and more robust. This leads to improved results as compared to the standard algorithms available out-of-the-box in state-of-the-art image registration toolboxes, both in terms of time as well as registration accuracy, which we empirically demonstrate on publicly available datasets.",
"authors": "Abhishek Nan, Matthew Tennant, Uriel Rubin, Nilanjan Ray",
"award": "",
"id": "64",
"or_id": "Q0Bm5e6dkW",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/nan20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/nan20.pdf",
"title": "DRMIME: Differentiable Mutual Information and Matrix Exponential for Multi-Resolution Image Registration",
"url": "papers/nan20.html",
"yt_full": "ryxGwCywVLY",
"yt_teaser": "RkWDByLlF5M"
},
"71": {
"abstract": "Exploiting learning algorithms under scarce data regimes is a limitation and a reality of the medical imaging field. In an attempt to mitigate the problem, we propose a data augmentation protocol based on generative adversarial networks. The networks are conditioned at a pixel-level (segmentation mask) and at a global-level information (acquisition environment or lesion type). Such conditioning provides immediate access to the image-label pairs while controlling class specific appearance of the synthesized images. To stimulate synthesis of the features relevant for the segmentation task, an additional passive player in a form of segmentor is introduced into the the adversarial game.\n We validate the approach on two medical datasets: BraTS, ISIC. By controlling the class distribution through injection of synthetic images into the training set we achieve control over the accuracy levels of the datasets' classes. ",
"authors": "Ahmad B Qasim, Ivan Ezhov, Suprosanna Shit, Oliver Schoppe, Johannes Paetzold, Anjany Sekuboyina, Florian Kofler, Jana Lipkova, Hongwei Li, Bjoern Menze",
"award": "",
"id": "71",
"or_id": "UHtZuvXHoA",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/qasim20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/qasim20.pdf",
"title": "Red-GAN: Attacking class imbalance via conditioned generation. Yet another medical imaging perspective.",
"url": "papers/qasim20.html",
"yt_full": "Dz81tWusvHk",
"yt_teaser": "HCP2X8z44VI"
},
"72": {
"abstract": "We present a deep learning strategy for contrast-agnostic semantic segmentation of unpreprocessed brain MRI scans, without requiring additional training or fine-tuning for new modalities. Classical Bayesian methods address this segmentation problem with unsupervised intensity models, but require significant computational resources. In contrast, learning-based methods can be fast at test time, but are sensitive to the data available at training. Our proposed learning method, SynthSeg, leverages a set of training segmentations (no intensity images required) to generate synthetic scans of widely varying contrasts on the fly during training. These scans are produced using the generative model of the classical Bayesian segmentation framework, with randomly sampled parameters for appearance, deformation, noise, and bias field. Because each mini-batch has a different synthetic contrast, the final network is not biased towards any specific MRI contrast. We comprehensively evaluate our approach on four datasets comprising over 1,000 subjects and four MR contrasts. The results show that our approach successfully segments every contrast in the data, performing slightly better than classical Bayesian segmentation, and three orders of magnitude faster. Moreover, even within the same type of MRI contrast, our strategy generalizes significantly better across datasets, compared to training using real images. Finally, we find that synthesizing a broad range of contrasts, even if unrealistic, increases the generalization of the neural network. Our code and model are open source at https://github.com/BBillot/SynthSeg.",
"authors": "Benjamin Billot, Douglas N. Greve, Koen Van Leemput, Bruce Fischl, Juan Eugenio Iglesias, Adrian V. Dalca",
"award": "",
"id": "72",
"or_id": "Qz2DgRQGlP",
"oral": "True",
"pmlr_url": "http://proceedings.mlr.press/v121/billot20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/billot20.pdf",
"title": "A learning strategy for contrast-agnostic MRI segmentation",
"url": "papers/billot20.html",
"yt_full": "Bfp3cILSKZg",
"yt_teaser": "ewKD9UqwUDk"
},
"73": {
"abstract": "Whether it be in a man-made machine or a biological system, form and function are often directly related. In the latter, however, this particular relationship is often unclear due to the intricate nature of biology. Here we developed a geometric deep learning model capable of exploiting the actual structure of the cortex to learn the complex relationship between brain function and anatomy from structural and functional MRI data. Our model was not only able to predict the functional organization of human visual cortex from anatomical properties alone, but it was also able to predict nuanced variations across individuals.",
"authors": "Fernanda L. Ribeiro, Steffen Bollmann, Alexander M. Puckett",
"award": "",
"id": "73",
"or_id": "Nw_trRFjPE",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/ribeiro20.pdf",
"title": "DeepRetinotopy: Predicting the Functional Organization of Human Visual Cortex from Structural MRI Data using Geometric Deep Learning",
"url": "papers/ribeiro20.html",
"yt_full": "DALdCiexUxo",
"yt_teaser": "kwHoPnCDvTk"
},
"89": {
"abstract": "Labelling large datasets for training high-capacity neural networks is a major obstacle to\n the development of deep learning-based medical imaging applications. Here we present a\n transformer-based network for magnetic resonance imaging (MRI) radiology report classification which automates this task by assigning image labels on the basis of free-text expert\n radiology reports. Our model\u2019s performance is comparable to that of an expert radiologist,\n and better than that of an expert physician, demonstrating the feasibility of this approach.\n We make code available online for researchers to label their own MRI datasets for medical\n imaging applications.",
"authors": "David Wood, Emily Guilhem, Antanas Montvila, Thomas Varsavsky, Martin Kiik, Juveria Siddiqui, Sina Kafiabadi, Naveen Gadapa, Aisha Al Busaidi, Matt Townend, Keena Patel, Gareth Barker, Sebastian Ourselin, Jeremy Lynch, James Cole, Tom Booth",
"award": "",
"id": "89",
"or_id": "9exoP7PDD3",
"oral": "True",
"pmlr_url": "http://proceedings.mlr.press/v121/wood20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/wood20.pdf",
"title": "Automated Labelling using an Attention model for Radiology reports of MRI scans (ALARM)",
"url": "papers/wood20.html",
"yt_full": "P1up3fFOh-g",
"yt_teaser": ""
},
"92": {
"abstract": "Multiple sclerosis lesion activity segmentation is the task of detecting new and enlarging lesions that appeared between a baseline and a follow-up brain MRI scan. While deep learning methods for single-scan lesion segmentation are common, deep learning approaches for lesion activity have only been proposed recently. Here, a two-path architecture processes two 3D MRI volumes from two time points. In this work, we investigate whether extending this problem to full 4D deep learning using a history of MRI volumes and thus an extended baseline can improve performance. For this purpose, we design a recurrent multi-encoder-decoder architecture for processing 4D data. We find that adding more temporal information is beneficial and our proposed architecture outperforms previous approaches with a lesion-wise true positive rate of 0.84 at a lesion-wise false positive rate of 0.19.",
"authors": "Nils Gessert, Marcel Bengs, Julia Kr\u00fcger, Roland Opfer, Ann-Christin Ostwaldt, Praveena Manogaran, Sven Schippling, Alexander Schlaefer",
"award": "",
"id": "92",
"or_id": "238UzYB1d9",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/gessert20.pdf",
"title": "4D Deep Learning for Multiple-Sclerosis Lesion Activity Segmentation",
"url": "papers/gessert20.html",
"yt_full": "Pu-8yWb1194",
"yt_teaser": "mISTl-A5EK8"
},
"93": {
"abstract": "Deep learning based image segmentation methods have achieved great success, even having human-level accuracy in some applications. However, due to the black box nature of deep learning, the best method may fail in some situations. Thus predicting segmentation quality without ground truth would be very crucial especially in clinical practice. Recently, people proposed to train neural networks to estimate the quality score by regression. Although it can achieve promising prediction accuracy, the network suffers robustness problem, e.g. it is vulnerable to adversarial attacks. In this paper, we propose to alleviate this problem by utilizing the difference between the input image and the reconstructed image, which is conditioned on the segmentation to be assessed, to lower the chance to overfit to the undesired image features from the original input image, and thus to increase the robustness. Results on ACDC17 dataset demonstrated our method is promising.",
"authors": "Leixin Zhou, Wenxiang Deng, Xiaodong Wu",
"award": "",
"id": "93",
"or_id": "nyhZXiaotm",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/zhou20.pdf",
"title": "Robust Image Segmentation Quality Assessment",
"url": "papers/zhou20.html",
"yt_full": "btP6vT6teT8",
"yt_teaser": "PgTTxpfsQvw"
},
"94": {
"abstract": "Nodule malignancy assessment is a complex, time-consuming and error-prone task. Current clinical practice requires measuring changes in size and density of the nodule at different time-points. State of the art solutions rely on 3D convolutional neural networks built on pulmonary nodules obtained from a single CT scan per patient. In this work, we propose a two-stream 3D convolutional neural network that predicts malignancy by jointly analyzing two pulmonary nodule volumes from the same patient taken at different time-points. Best results achieve 77% of F1-score in test with an increment of 9% and 12% of F1-score with respect to the same network trained with images from a single time-point.",
"authors": "Xavier Rafael-Palou, Anton Aubanell, Ilaria Bonavita, Mario Ceresa, Gemma Piella, Vicent Ribas, Miguel \u00c1ngel Gonz\u00e1lez Ballester",
"award": "",
"id": "94",
"or_id": "D1jTt_FOPY",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/rafael-palou20.pdf",
"title": "Pulmonary Nodule Malignancy Classification Using its Temporal Evolution with Two-Stream 3D Convolutional Neural Networks",
"url": "papers/rafael-palou20.html",
"yt_full": "buuH4as0rlk",
"yt_teaser": "ifUQ8CBSUkQ"
},
"95": {
"abstract": "With the increasing adoption of machine learning tools like neural networks across several domains, interesting connections and comparisons to concepts from other domains are coming to light. In this work, we focus on the class of Tensor Networks, which has been a work horse for physicists in the last two decades to analyse quantum many-body systems. Building on the recent interest in tensor networks for machine learning, we extend the Matrix Product State tensor networks (which can be interpreted as linear classifiers operating in exponentially high dimensional spaces) to be useful in medical image analysis tasks. We focus on classification problems as a first step where we motivate the use of tensor networks and propose adaptions for 2D images using classical image domain concepts such as local orderlessness of images. With the proposed locally orderless tensor network model (LoTeNet), we show that tensor networks are capable of attaining performance that is comparable to state-of-the-art deep learning methods. We evaluate the model on two publicly available medical imaging datasets and show performance improvements with fewer model hyperparameters and lesser computational resources compared to relevant baseline methods.",
"authors": "Raghavendra Selvan, Erik B Dam",
"award": "Runner-up for best paper Award",
"id": "95",
"or_id": "jjk6bxk07G",
"oral": "True",
"pmlr_url": "http://proceedings.mlr.press/v121/selvan20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/selvan20.pdf",
"title": "Tensor Networks for Medical Image Classification",
"url": "papers/selvan20.html",
"yt_full": "CpBJVULSGiY",
"yt_teaser": "w0YgIa2SxWk"
},
"96": {
"abstract": "We focus on the problem of training convolutional neural networks on gigapixel histopathology images to predict image-level targets. For this purpose, we extend Neural Image Compression (NIC), an image compression framework that reduces the dimensionality of these images using an encoder network trained unsupervisedly. We propose to train this encoder using supervised multitask learning (MTL) instead. We applied the proposed MTL NIC to two histopathology datasets and three tasks. First, we obtained state-of-the-art results in the Tumor Proliferation Assessment Challenge of 2016 (TUPAC16). Second, we successfully classified histopathological growth patterns in images with colorectal liver metastasis (CLM). Third, we predicted patient risk of death by learning directly from overall survival in the same CLM data. Our experimental results suggest that the representations learned by the MTL objective are: (1) highly specific, due to the supervised training signal, and (2) transferable, since the same features perform well across different tasks. Additionally, we trained multiple encoders with different training objectives, e.g. unsupervised and variants of MTL, and observed a positive correlation between the number of tasks in MTL and the system performance on the TUPAC16 dataset.",
"authors": "David Tellez, Diederik Hoppener, Cornelis Verhoef, Dirk Grunhagen, Pieter Nierop, Michal Drozdzal, Jeroen van der Laak, Francesco Ciompi",
"award": "Runner-up for best paper Award",
"id": "96",
"or_id": "oepOBj_A7E",
"oral": "True",
"pmlr_url": "http://proceedings.mlr.press/v121/tellez20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/tellez20.pdf",
"title": "Extending Unsupervised Neural Image Compression With Supervised Multitask Learning",
"url": "papers/tellez20.html",
"yt_full": "w48yYvYixCk",
"yt_teaser": "GhW1qVGVit8"
},
"99": {
"abstract": "The scarcity of labeled data often limits the application of deep learning to medical image segmentation. Semi-supervised learning helps overcome this limitation by leveraging unlabeled images to guide the learning process. In this paper, we propose using a clustering loss based on mutual information that explicitly enforces prediction consistency between nearby pixels in unlabeled images, and for random perturbation of these images, while imposing the network to predict the correct labels for annotated images. Since mutual information does not require a strict ordering of clusters in two different cluster assignments, we propose to incorporate another consistency regularization loss which forces the alignment of class probabilities at each pixel of perturbed unlabeled images. We evaluate the method on three challenging publicly-available medical datasets for image segmentation. Experimental results show our method to outperform recently-proposed approaches for semi-supervised and yield a performance comparable to fully-supervised training.",
"authors": "Jizong Peng, Marco Pedersoli, Christian Desrosiers",
"award": "",
"id": "99",
"or_id": "iunvffXgPm",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/peng20b.html",
"schedule": "",
"short": "False",
"slides": "/slides/peng20a.pdf",
"title": "Mutual information deep regularization for semi-supervised segmentation",
"url": "papers/peng20a.html",
"yt_full": "tLUqomGRmCU",
"yt_teaser": ""
},
"102": {
"abstract": "State-of-the-art deep learning based methods have achieved remarkable performance on medical image segmentation. Their applications in the clinical setting are, however, limited due to the lack of trustworthiness and reliability. Selective image segmentation has been proposed to address this issue by letting a DNN model process instances with high confidence while referring difficult ones with high uncertainty to experienced radiologists. As such, the model performance is only affected by the predictions on the high confidence subset rather than the whole dataset. Existing selective segmentation methods, however, ignore this unique property of selective segmentation and train their DNN models by optimizing accuracy on the entire dataset. Motivated by such a discrepancy, we present a novel method in this paper that considers such uncertainty in the training process to maximize the accuracy on the confident subset rather than the accuracy on the whole dataset. Experimental results using the whole heart and great vessel segmentation and gland segmentation show that such a training scheme can significantly improve the performance of selective segmentation. ",
"authors": "Yukun Ding, Jinglan Liu, Xiaowei Xu, Meiping Huang, Jian Zhuang, Jinjun Xiong, Yiyu Shi",
"award": "",
"id": "102",
"or_id": "F1MIJCqX2J",
"oral": "True",
"pmlr_url": "http://proceedings.mlr.press/v121/ding20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/ding20.pdf",
"title": "Uncertainty-Aware Training of Neural Networks for Selective Medical Image Segmentation",
"url": "papers/ding20.html",
"yt_full": "iHU9eQV_hiI",
"yt_teaser": ""
},
"103": {
"abstract": "Deep neural networks are powerful tools for biomedical image segmentation. These models are often trained with heavy supervision, relying on pairs of images and corresponding voxel-level labels. However, obtaining segmentations of anatomical regions on a large number of cases can be prohibitively expensive. Thus there is a strong need for deep learning-based segmentation tools that do not require heavy supervision and can continuously adapt. In this paper, we propose a novel perspective of segmentation as a discrete representation learning problem, and present a variational autoencoder segmentation strategy that is flexible and adaptive. Our method, called Segmentation Auto-Encoder (SAE), leverages all available unlabeled scans and merely requires a segmentation prior, which can be a single unpaired segmentation image. In experiments, we apply SAE to brain MRI scans. Our results show that SAE can produce good quality segmentations, particularly when the prior is good. We demonstrate that a Markov Random Field prior can yield significantly better results than a spatially independent prior. Our code is freely available at https://github.com/evanmy/sae. ",
"authors": "Evan M. Yu, Juan Eugenio Iglesias, Adrian V. Dalca, Mert R. Sabuncu",
"award": "",
"id": "103",
"or_id": "J1-4vNudWo",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/yu20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/yu20.pdf",
"title": "An Auto-Encoder Strategy for Adaptive Image Segmentation",
"url": "papers/yu20.html",
"yt_full": "6c1jNm8qC5A",
"yt_teaser": "WYpKmCzMuH8"
},
"104": {
"abstract": "Ovarian cancer is the most lethal cancer of the female reproductive organs. There are $5$ major histological subtypes of epithelial ovarian cancer, each with distinct morphological, genetic, and clinical features. Currently, these histotypes are determined by a pathologist's microscopic examination of tumor whole-slide images (WSI). This process has been hampered by poor inter-observer agreement (Cohen\u2019s kappa $0.54$-$0.67$). We utilized a two-stage deep transfer learning algorithm based on convolutional neural networks (CNN) and progressive resizing for automatic classification of epithelial ovarian carcinoma WSIs. The proposed algorithm achieved a mean accuracy of $87.54\\%$ and Cohen's kappa of $0.8106$ in the slide-level classification of $305$ WSIs; performing better than a standard CNN and pathologists without gynecology-specific training. ",
"authors": "Yiping Wang, David Farnell, Hossein Farahani, Mitchell Nursey, Basile Tessier-Cloutier, Steven J.M. Jones, David G. Huntsman, C. Blake Gilks, Ali Bashashati",
"award": "",
"id": "104",
"or_id": "VXdQD8B307",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/wang20b.pdf",
"title": "Classification of Epithelial Ovarian Carcinoma Whole-Slide Pathology Images Using Deep Transfer Learning",
"url": "papers/wang20b.html",
"yt_full": "nS_74yeosKo",
"yt_teaser": "I2JsLBARtmo"
},
"107": {
"abstract": "Saliency maps have become a widely used method to assess which areas of the input image are most pertinent to the prediction of a trained neural network. However, in the context of medical imaging, there is no study to our knowledge that has examined the efficacy of these techniques and quantified them using overlap with ground truth bounding boxes. In this work, we explored the credibility of the various existing saliency map methods on the RSNA Pneumonia dataset. We found that GradCAM was the most sensitive to model parameter and label randomization, and was highly agnostic to model architecture.",
"authors": "Nishanth Thumbavanam Arun, Nathan Gaw, Praveer Singh, Ken Chang, Katharina Viktoria Hoebel, Jay Patel, Mishka Gidwani, Jayashree Kalpathy-Cramer",
"award": "",
"id": "107",
"or_id": "02X3kfP6W4",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/arun20.pdf",
"title": "Assessing the validity of saliency maps for abnormality localization in medical imaging",
"url": "papers/arun20.html",
"yt_full": "wrw_i134jl0",
"yt_teaser": "8V99wNHzJbQ"
},
"109": {
"abstract": "Tracking and segmentation of biological cells in video sequences is a challenging problem, especially due to the similarity of the cells and high levels of inherent noise. Most machine learning-based approaches lack robustness and suffer from sensitivity to less prominent events such as mitosis, apoptosis and cell collisions. Due to the large variance in medical image characteristics, most approaches are dataset-specific and do not generalise well on other datasets.\n\n In this paper, we propose a simple end-to-end cascade neural architecture able to model the movement behaviour of biological cells and predict collision and mitosis events. Our approach uses U-Net for an initial segmentation which is then improved through processing by a siamese tracker capable of matching each cell along the temporal axis. By facilitating the re-segmentation of collided and mitotic cells, our method demonstrates its capability to handle volatile trajectories and unpredictable cell locations while being invariant to cell morphology. We demonstrate that our tracking approach achieves state-of-the-art results on PhC-C2DL-PSC and Fluo-N2DH-SIM+ datasets and ranks second on the DIC-C2DH-HeLa dataset of the cell tracking challenge benchmarks. ",
"authors": "Andreas Panteli, Deepak K. Gupta, Nathan de Bruin, Efstratios Gavves",
"award": "",
"id": "109",
"or_id": "V3ZrDLgNgu",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/panteli20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/panteli20.pdf",
"title": "Siamese Tracking of Cell Behaviour Patterns",
"url": "papers/panteli20.html",
"yt_full": "AnEDZ76eOmY",
"yt_teaser": "48HXH_WYi6k"
},
"114": {
"abstract": "Training deep convolution neural network requires a large amount of data to obtain good performance and generalisable results. Transfer learning approaches from datasets such as ImageNet had become important in increasing accuracy and lowering training samples required. However, as of now, there has not been a popular dataset for training 3D volumetric medical images. This is mainly due to the time and expert knowledge required to accurately annotate medical images. In this study, we present a method in extracting labels from DICOM metadata that information on the appearance of the scans to train a medical domain 3D convolution neural network. The labels include imaging modalities and sequences, patient orientation and view, presence of contrast agent, scan target and coverage, and slice spacing. We applied our method and extracted labels from a large amount of cancer imaging dataset from TCIA to train a medical domain 3D deep convolution neural network. We evaluated the effectiveness of using our proposed network in transfer learning a liver segmentation task and found that our network achieved superior segmentation performance (DICE=90.0%) compared to training from scratch (DICE=41.8%). Our proposed network shows promising results to be used as a backbone network for transfer learning to another task. Our approach along with the utilising our network, can potentially be used to extract features from large-scale unlabelled DICOM datasets.",
"authors": "Richard Du, Varut Vardhanabhuti",
"award": "",
"id": "114",
"or_id": "CCbuElJreP",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/du20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/du20.pdf",
"title": "3D-RADNet: Extracting labels from DICOM metadata for training general medical domain deep 3D convolution neural networks",
"url": "papers/du20.html",
"yt_full": "b0XF-PLmmnM",
"yt_teaser": "6pRjwuIsedk"
},
"116": {
"abstract": "In the WHO glioma classification guidelines grade, IDH mutation and 1p19q co-deletion play a central role as they are important markers for prognosis and optimal therapy planning. Therefore, we propose a fully automatic, MRI based, 3D pipeline for glioma segmentation and classification. The designed segmentation network was a 3D U-Net achieving an average whole tumor dice score of 90%. After segmentation, the 3D tumor ROI is extracted and fed into the multi-task classification network. The network was trained and evaluated on a large heterogeneous dataset of 628 patients, collected from The Cancer Imaging Archive and BraTS 2019 databases. Additionally, the network was validated on an independent dataset of 110 patients retrospectively acquired at the Ghent University Hospital (GUH). Classification AUC scores are 0.93, 0.94 and 0.82 on the TCIA test data and 0.94, 0.86 and 0.87 on the GUH data for grade, IDH and 1p19q status respectively. ",
"authors": "Milan Decuyper, Stijn Bonte, Karel Deblaere, Roel Van Holen",
"award": "",
"id": "116",
"or_id": "J5iep2t90F",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/decuyper20.pdf",
"title": "Automated MRI based pipeline for glioma segmentation and prediction of grade, IDH mutation and 1p19q co-deletion",
"url": "papers/decuyper20.html",
"yt_full": "vos32kY5pak",
"yt_teaser": "HmU_h8GaUQM"
},
"117": {
"abstract": "Magnetic Resonance Imaging (MRI) is a widely-accepted imaging technique for knee injury analysis. Its advantage of capturing knee structure in three dimensions makes it the ideal tool for radiologists to locate potential tears in the knee. In order to better confront the ever growing workload of musculoskeletal (MSK) radiologists, automated tools for patients' triage are becoming a real need, reducing delays in the reading of pathological cases. In this work, we present the Efficiently-Layered Network (ELNet), a convolutional neural network (CNN) architecture optimized for the task of initial knee MRI diagnosis for triage. Unlike past approaches, we train ELNet from scratch instead of using a transfer-learning approach. The proposed method is validated quantitatively and qualitatively, and compares favorably against state-of-the-art MRNet while using a single imaging stack (axial or coronal) as input. Additionally, we demonstrate our model's capability to locate tears in the knee despite the absence of localization information during training. Lastly, the proposed model is extremely lightweight ($<$ 1MB) and therefore easy to train and deploy in real clinical settings.",
"authors": "Chen-Han Tsai, Nahum Kiryati, Eli Konen, Iris Eshed, Arnaldo Mayer",
"award": "",
"id": "117",
"or_id": "B_NG9y_wqU",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/tsai20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/tsai20.pdf",
"title": "Knee Injury Detection using MRI with Efficiently-Layered Network (ELNet)",
"url": "papers/tsai20.html",
"yt_full": "ucWYdEJ545k",
"yt_teaser": "8nO-E_2aNcE"
},
"118": {
"abstract": "In this work, we propose a Variational Autoencoder (VAE) - Generative Adversarial Networks (GAN) model that can produce highly realistic MRI together with its pixel accurate groundtruth for the application of cine-MR image cardiac segmentation. On one side of our model is a Variational Autoencoder (VAE) trained to learn the latent representations of cardiac shapes. On the other side is a GAN that uses \u201dSPatially-Adaptive (DE)Normalization\u201d (SPADE) modules to generate realistic MR images tailored to a given anatomical map. At test time, the sampling of the VAE latent space allows to generate an arbitrary large number of cardiac shapes, which are fed to the GAN that subsequently generates MR images whose cardiac structure fits that of the cardiac shapes. In other words, our system can generate a large volume of realistic yet labeled cardiac MR images. We show that segmentation with CNNs trained with our synthetic annotated images gets competitive results compared to traditional techniques.\n We also show that combining data augmentation with our GAN-generated images lead to an improvement in the Dice score of up to 12 percent while allowing for better generalization capabilities on other datasets.",
"authors": "Youssef Skandarani, Nathan Painchaud, Pierre-Marc Jodoin, Alain Lalande",
"award": "",
"id": "118",
"or_id": "f9Pl3Qj3_Q",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/skandarani20.pdf",
"title": "On the effectiveness of GAN generated cardiac MRIs for segmentation",
"url": "papers/skandarani20.html",
"yt_full": "iteI5M0BcFY",
"yt_teaser": "lIVWmPNFG1E"
},
"126": {
"abstract": "The process of fertilizing a human egg outside the body in order to help those suffering from infertility to conceive is known as in vitro fertilization (IVF). Despite being the most effective method of assisted reproductive technology (ART), the average success rate of IVF is a mere 20-40%. One step that is critical to the success of the procedure is selecting which embryo to transfer to the patient, a process typically conducted manually and without any universally accepted and standardized criteria. In this paper we describe a novel data-driven system trained to directly predict embryo implantation probability from embryogenesis time-lapse imaging videos. Using retrospectively collected videos from 272 embryos, we demonstrate that, when compared to an external panel of embryologists, our algorithm results in a 12% increase of positive predictive value and a 29% increase of negative predictive value. ",
"authors": "David H. Silver, Martin Feder, Yael Gold-Zamir, Avital L. Polsky, Shahar Rosentraub, Efrat Shachor, Adi Weinberger, Pavlo Mazur, Valery D. Zukin, Alex M. Bronstein",
"award": "",
"id": "126",
"or_id": "TujK1uTkTP",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/silver20.pdf",
"title": "Data-Driven Prediction of Embryo Implantation Probability Using IVF Time-lapse Imaging",
"url": "papers/silver20.html",
"yt_full": "pN5R4EbkbKU",
"yt_teaser": "Jg8Aq4BuXI8"
},
"128": {
"abstract": "Robust localization of organs in computed tomography scans is a constant pre-processing requirement for organ-specific image retrieval, radiotherapy planning, and interventional image analysis. In contrast to current solutions based on exhaustive search or region proposals, which require large amounts of annotated data, we propose a deep reinforcement learning approach for organ localization in CT. In this work, an artificial agent is actively self-taught to localize organs in CT by learning from its asserts and mistakes. Within the context of reinforcement learning, we propose a novel set of actions tailored for organ localization in CT. Our method can use as a plug-and-play module for localizing any organ of interest. We evaluate the proposed solution on the public VISCERAL dataset containing CT scans with varying fields of view and multiple organs. We achieved an overall intersection over union of 0.63, an absolute median wall distance of 2.25 mm and a median distance between centroids of 3.65 mm.",
"authors": "Fernando Navarro, Anjany Sekuboyina, Diana Waldmannstetter, Jan C. Peeken, Stephanie E. Combs, Bjoern H. Menze",
"award": "",
"id": "128",
"or_id": "0vDeD2UD0S",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/navarro20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/navarro20.pdf",
"title": "Deep Reinforcement Learning for Organ Localization in CT",
"url": "papers/navarro20.html",
"yt_full": "tQVNgKCWs9I",
"yt_teaser": "Rk6gCh7eYkU"
},
"131": {
"abstract": "Multimodal image registration is a very challenging problem for deep learning approaches. Most current work focuses on either supervised learning that requires labelled training scans and may yield models that bias towards annotated structures or unsupervised approaches that are based on hand-crafted similarity metrics and may therefore not outperform their classical non-trained counterparts. We believe that unsupervised domain adaptation can be beneficial in overcoming the current limitations for multimodal registration, where good metrics are hard to define.\n Domain adaptation has so far been mainly limited to classification problems. We propose the first use of unsupervised domain adaptation for discrete multimodal registration. Based on a source domain for which quantised displacement labels are available as supervision, we transfer the output distribution of the network to better resemble the target domain (other modality) using classifier discrepancies. To improve upon the sliced Wasserstein metric for 2D histograms, we present a novel approximation that projects predictions into 1D and computes the L1 distance of their cumulative sums. Our proof-of-concept demonstrates the applicability of domain transfer from mono- to multimodal 2D registration of canine MRI scans and improves the registration accuracy from 33% (using sliced Wasserstein) to 44%.",
"authors": "Mattias P Heinrich, Lasse Hansen",
"award": "",
"id": "131",
"or_id": "wbZM-DcJB9",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/heinrich20.pdf",
"title": "Unsupervised learning of multimodal image registration using domain adaptation with projected Earth Mover\u2019s discrepancies",
"url": "papers/heinrich20.html",
"yt_full": "yfj_hAg89rg",
"yt_teaser": "MvcUYnalf4U"
},
"133": {
"abstract": "Current deep learning methods are based on the repeated, expensive application of convolutions with parameter-intensive weight matrices. In this work, we present a novel concept that enables the application of differentiable random ferns in end-to-end networks. It can then be used as multiplication-free convolutional layer alternative in deep network architectures. Our experiments on the binary classification task of the TUPAC'16 challenge demonstrate improved results over the state-of-the-art binary XNOR net and only slightly worse performance than its 2x more parameter intensive floating point CNN counterpart. ",
"authors": "Max Blendowski, Mattias P. Heinrich",
"award": "",
"id": "133",
"or_id": "EiT7GQAj-T",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/blendowski20.pdf",
"title": "Learning to map between ferns with differentiable binary embedding networks",
"url": "papers/blendowski20.html",
"yt_full": "vetTlos9ug0",
"yt_teaser": "a0266OkLfhw"
},
"135": {
"abstract": "Cardiac left ventricular (LV) segmentation from short-axis MRI acquired 10 minutes after the injection of a contrast agent (LGE-MRI) is a necessary step in the processing allowing the identification and diagnosis of cardiac diseases such as myocardial infarction. However, this segmentation is challenging due to high variability across subjects and the potential lack of contrast between structures. Then, the main objective of this work is to develop an accurate automatic segmentation method based on deep learning models for the myocardial borders on LGE-MRI. To this end, 2.5 D residual neural network integrated with a squeeze and excitation blocks in encoder side with specialized convolutional has been proposed. Late fusion has been used to merge the output of the best trained proposed models from a different set of hyperparameters. A total number of 320 exams (with a mean number of 6 slices per exam) were used for training and 28 exams used for testing. The performance analysis of the proposed ensemble model in the basal and middle slices was similar as compared to intra-observer study and slightly lower at apical slices. The overall Dice score was 82.01% by our proposed method as compared to Dice score of 83.22% obtained from the intra observer study. The proposed model could be used for the automatic segmentation of myocardial border that is a very important step for accurate quantification of no-reflow, myocardial infarction, myocarditis, and hypertrophic cardiomyopathy, among others.",
"authors": "Abdul Qayyum, Alain Lalande, Thomas Decourselle, Thibaut Pommier, Alexandre Cochet, Fabrice Meriaudeau",
"award": "",
"id": "135",
"or_id": "4v2lR3Zvsw",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/qayyum20.pdf",
"title": "Segmentation of the Myocardium on Late-Gadolinium Enhanced MRI based on 2.5 D Residual Squeeze and Excitation Deep Learning Model",
"url": "papers/qayyum20.html",
"yt_full": "hdNBYfsuJSE",
"yt_teaser": "Gl1ioiS7IEE"
},
"136": {
"abstract": "We developed an automated system based on deep neural networks for fast and sensitive 3D image segmentation of cortical gray matter from fetal brain MRI. The lack of extensive/publicly available annotations presented a key challenge, as large amounts of labeled data are typically required for training sensitive models with deep learning. To address this, we: (i) generated preliminary tissue labels using the Draw-EM algorithm, which uses Expectation-Maximization and was originally designed for tissue segmentation in the neonatal domain; and (ii) employed a human-in-the-loop approach, whereby an expert fetal imaging annotator assessed and refined the performance of the model. By using a hybrid approach that combined automatically generated labels with manual refinements by an expert, we amplified the utility of ground truth annotations while immensely reducing their cost (283 slices). The deep learning system was developed, refined, and validated on 249 3D T2-weighted scans obtained from the Developing Human Connectome Project's fetal cohort, acquired at 3T. Analysis of the system showed that it is invariant to gestational age at scan, as it generalized well to a wide age range (21 \u2013 38 weeks) despite variations in cortical morphology and intensity across the fetal distribution. It was also found to be invariant to intensities in regions surrounding the brain (amniotic fluid), which often present a major obstacle to the processing of neuroimaging data in the fetal domain. ",
"authors": "Ahmed E. Fetit, Amir Alansary, Lucilio Cordero-Grande, John Cupitt, Alice B. Davidson, A. David Edwards, Joseph V. Hajnal, Emer Hughes, Konstantinos Kamnitsas, Vanessa Kyriakopoulou, Antonios Makropoulos, Prachi A. Patkee, Anthony N. Price, Mary A. Rutherford, Daniel Rueckert",
"award": "",
"id": "136",
"or_id": "VtVIlHSc0",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/fetit20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/fetit20b.pdf",
"title": "A deep learning approach to segmentation of the developing cortex in fetal brain MRI with minimal manual labeling",
"url": "papers/fetit20b.html",
"yt_full": "AahUnVAA2VU",
"yt_teaser": "CJIAB_ryWGs"
},
"137": {
"abstract": "We work on the breast imaging malignancy segmentation task while focusing on the train- ing process instead of network complexity. We designed a training process based on a modified U-Net, increasing the overall segmentation performances by using both, benign and malignant data for training. Our approach makes use of only a small amount of anno- tated data and relies on transfer learning from a self-supervised reconstruction task, and favors explainability.",
"authors": "Mickael Tardy, Diana Mateus",
"award": "",
"id": "137",
"or_id": "vVsWe9-s0G",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/tardy20.pdf",
"title": "Improving Mammography Malignancy Segmentation by Designing the Training Process",
"url": "papers/tardy20.html",
"yt_full": "mHJ3UmduXNw",
"yt_teaser": "w8W_NtJ2Rjg"
},
"140": {
"abstract": "Successful clinical implementation of deep learning in medical imaging depends, in part, on the reliability of the predictions. Specifically, the system should be accurate for classes seen during training while providing calibrated estimates of uncertainty for abnormalities and unseen classes. To efficiently estimate predictive uncertainty, we propose the use of multi-head convolutional neural networks (M-heads). We compare its performance to related and more prevalent approaches, such as deep ensembles, on the task of out-of-distribution (OOD) detection. To this end, we evaluate models, trained to discriminate normal lymph node tissue from breast cancer metastases, on lymph nodes containing lymphoma. We show the ability to discriminate between the in-distribution lymph node tissue and lymphoma by evaluating the AUROC based on the uncertainty signal. Here, the best performing multi-head CNN (91.7) outperforms both Monte Carlo dropout (86.5) and deep ensembles (86.8). Furthermore, we show that the meta-loss function of M-heads improves OOD detection in terms of AUROC from 87.4 to 88.7.",
"authors": "Jasper Linmans, Jeroen van der Laak, Geert Litjens",
"award": "",
"id": "140",
"or_id": "hRwB2BTRNu",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/linmans20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/linmans20.pdf",
"title": "Efficient Out-of-Distribution Detection in Digital Pathology Using Multi-Head Convolutional Neural Networks",
"url": "papers/linmans20.html",
"yt_full": "P_fTmR2Fwp8",
"yt_teaser": "aK8XWnWOKaE"
},
"145": {
"abstract": "Medical image classification performance worsens in multi-domain datasets, caused by radiological image differences across institutions, scanner manufacturer, model and operator. Deep learning is well-suited for learning image features with priors encoded as constraints during the training process. In this work, we apply a ResNeXt classification network augmented with an FCN preprocessor subnetwork to a public TCIA head and neck cancer dataset. The training goal is survival prediction of radiotherapy cases based on pre-treatment FDG-PET/CT scans, acquired across 4 different hospitals. We show that the preprocessor sub-network acts as a embedding normalizer and improves over state-of-the-art results of 70% AUC to 76%.",
"authors": "William Le, Francisco Perdig\u00f3n Romero, Samuel Kadoury",
"award": "",
"id": "145",
"or_id": "JojEzQ3E5n",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/le20.pdf",
"title": "A Fully Convolutional Normalization Approach of Head and Neck Cancer Outcome Prediction",
"url": "papers/le20.html",
"yt_full": "JXM-25mYIh0",
"yt_teaser": "OD-lmWkCr4A"
},
"146": {
"abstract": "Organ segmentation in CT volumes is an important pre-processing step in many computer assisted intervention and diagnosis methods. In recent years, convolutional neural networks have dominated the state of the art in this task. However, since this problem presents a challenging environment due to high variability in the organ's shape and similarity between tissues, the generation of false negative and false positive regions in the output segmentation is a common issue. Recent works have shown that the uncertainty analysis of the model can provide us with useful information about potential errors in the segmentation. In this context, we proposed a segmentation refinement method based on uncertainty analysis and graph convolutional networks. We employ the uncertainty levels of the convolutional network in a particular input volume to formulate a semi-supervised graph learning problem that is solved by training a graph convolutional network. To test our method we refine the initial output of a 2D U-Net. We validate our framework with the NIH pancreas dataset and the spleen dataset of the medical segmentation decathlon. We show that our method outperforms the state-of-the art CRF refinement method by improving the dice score by 1% for the pancreas and 2% for spleen, with respect to the original U-Net's prediction. Finally, we discuss the results and current limitations of the model for future work in this research direction. For reproducibility purposes, we make our code publicly available",
"authors": "Roger D. Soberanis-Mukul, Nassir Navab, Shadi Albarqouni",
"award": "",
"id": "146",
"or_id": "UUie86nf5B",
"oral": "True",
"pmlr_url": "http://proceedings.mlr.press/v121/soberanis-mukul20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/soberanis--mukul20.pdf",
"title": "Uncertainty-based Graph Convolutional Networks for Organ Segmentation Refinement",
"url": "papers/soberanis--mukul20.html",
"yt_full": "VwF0KeALAFU",
"yt_teaser": ""
},
"150": {
"abstract": "We apply Generative Adversarial Networks (GANs) to the domain of digital pathology. Current machine learning research for digital pathology focuses on diagnosis, but we suggest a different approach and advocate that generative models could drive forward the understanding of morphological characteristics of cancer tissue. In this paper, we develop a framework which allows GANs to capture key tissue features and uses these characteristics to give structure to its latent space. To this end, we trained our model on 249K H&E breast cancer tissue images.\n We show that our model generates high quality images, with a Frechet Inception Distance (FID) of 16.65. We additionally assess the quality of the images with cancer tissue characteristics (e.g. count of cancer, lymphocytes, or stromal cells), using quantitative information to calculate the FID and showing consistent performance of 9.86. Additionally, the latent space of our model shows an interpretable structure and allows semantic vector operations that translate into tissue feature transformations. Furthermore, ratings from two expert pathologists found no significant difference between our generated tissue images from real ones.",
"authors": "Adalberto Claudio Quiros, Roderick Murray-Smith, Ke Yuan",
"award": "",
"id": "150",
"or_id": "CwgSEEQkad",
"oral": "False",
"pmlr_url": "http://proceedings.mlr.press/v121/quiros20a.html",
"schedule": "",
"short": "False",
"slides": "/slides/quiros20.pdf",
"title": "PathologyGAN: Learning deep representations of cancer tissue",
"url": "papers/quiros20.html",
"yt_full": "RlBo4lAWtvM",
"yt_teaser": "wGx78e1mHQQ"
},
"155": {
"abstract": "Though, deep learning based medical image registration is currently starting to show promising advances, often, it still fells behind conventional frameworks in terms of reg- istration accuracy. This is especially true for applications where large deformations exist, such as registration of interpatient abdominal MRI or inhale-to-exhale CT lung registra- tion. Most current works use U-Net-like architectures to predict dense displacement fields from the input images in different supervised and unsupervised settings. We believe that the U-Net architecture itself to some level limits the ability to predict large deformations (even when using multilevel strategies) and therefore propose a novel approach, where the input images are mapped into a displacement space and final registrations are reconstructed from this embedding. Experiments on inhale-to-exhale CT lung registration demonstrate the ability of our architecture to predict large deformations in a single forward path through our network (leading to errors below 2 mm).",
"authors": "Lasse Hansen, Mattias P. Heinrich",
"award": "",
"id": "155",
"or_id": "kPBUZluVq",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/hansen20.pdf",
"title": "Tackling the Problem of Large Deformations in Deep Learning Based Medical Image Registration Using Displacement Embeddings",
"url": "papers/hansen20.html",
"yt_full": "f3oAOc7hNkI",
"yt_teaser": "2uJ_H0G1f-0"
},
"156": {
"abstract": "Forecasting motion of a specific target object is a common problem for surgical interventions, e.g. for localization of a target region, guidance for surgical interventions, or motion compensation. Optical coherence tomography (OCT) is an imaging modality with a high spatial and temporal resolution. Recently, deep learning methods have shown promising performance for OCT-based motion estimation based on two volumetric images. We extend this approach and investigate whether using a time series of volumes enables motion forecasting. We propose 4D spatio-temporal deep learning for end-to-end motion forecasting and estimation using a stream of OCT volumes. We design and evaluate five different 3D and 4D deep learning methods using a tissue data set. Our best performing 4D method achieves motion forecasting with an overall average correlation coefficient of 97.41%, while also improving motion estimation performance by a factor of 2.5 compared to a previous 3D approach. ",
"authors": "Marcel Bengs, Nils Gessert, Alexander Schlaefer",
"award": "",
"id": "156",
"or_id": "WVd56kgRV",
"oral": "False",
"pmlr_url": "",
"schedule": "",
"short": "True",
"slides": "/slides/bengs20.pdf",
"title": "A Deep Learning Approach for Motion Forecasting Using 4D OCT Data",
"url": "papers/bengs20.html",
"yt_full": "IBQUuRt4X9o",
"yt_teaser": "32BZ5eOHzTk"
},
"157": {
"abstract": "Large labeled datasets are an important precondition for deep learning models to achieve state-of-the-art results in computer vision tasks. In the medical imaging domain, privacy concerns have limited the rate of adoption of artificial intelligence methodologies into clinical practice. To alleviate such concerns, and increase comfort levels while sharing and storing surgical video data, we propose a high accuracy method for rapid removal and anonymization of out-of-body and non-relevant surgery segments. Training a deep model to detect out-of-body and non-relevant segments in surgical videos requires suitable labeling. Since annotating surgical videos with per-second relevancy labeling is a tedious task, our proposed framework initiates the learning process from a weakly labeled noisy dataset and iteratively applies Semi-Supervised Learning (SSL) to re-annotate the training data samples. Evaluating our model, on an independent test set, shows a mean detection accuracy of above 97% after several training-annotating iterations. Since our final goal is achieving out-of-body segments detection for anonymization, we evaluate our ability to detect these segments at a high demanding recall of 97%, which leads to a precision of 83.5%. We believe this approach can be applied to similar related medical problems, in which only a coarse set of relevancy labels exists, currently limiting the possibility for supervision training.",
"authors": "Maya Zohar, Omri Bar, Daniel Neimark, Gregory D. Hager, Dotan Asselmann",
"award": "",
"id": "157",
"or_id": "k-ANsPQJxY",
"oral": "False",