The resulting produced by the neural network.
The number usually denotes a specific test case, scene, or figure number referenced within the study. This paper explores the vulnerability of deep learning-based image segmentation models (like those used in autonomous driving) to adversarial patches—small, intentionally designed images that can cause a model to misclassify specific objects or entire regions of a scene. Context of the Paper apns-218.mp4
You can often find these supplementary videos on platforms like arXiv (under the "Ancillary files" section) or the researchers' project GitHub repositories. The resulting produced by the neural network