The double JPEG compression detection has received much attention in recent years due to its applicability as a forensic tool for the most widely used JPEG file format. Existing state-of-the-art CNN-based methods either use histograms of all the frequencies or rely on heuristics to select histograms of specific low frequencies to classify single and double compressed images. However, even amidst lower frequencies of double compressed images/patches, histograms of all the frequencies do not have distinguishable features to separate them from single compressed images. This paper directly extracts the quantized DCT coefficients from the JPEG images without decompressing them in the pixel domain, obtains all AC frequencies' histograms, uses a module based on $1\times 1$ depth-wise convolutions to learn the inherent relation between each histogram and corresponding q-factor, and utilizes a tailor-made BiLSTM network for selectively encoding these feature vector sequences. The proposed system outperforms several baseline methods on a relatively large and diverse publicly available dataset of single and double compressed patches. Another essential aspect of any single vs. double JPEG compression detection system is handling the scenario where test patches are compressed with entirely different quantization matrices (Q-matrices) than those used while training; different camera manufacturers and image processing software generally utilize their customized quantization matrices. A set of extensive experiments shows that the proposed system trained on a single dataset generalizes well on other datasets compressed with completely unseen quantization matrices and outperforms the state-of-the-art methods in both seen and unseen quantization matrices scenarios.
Impact craters are formed as a result of continuous impacts on the surface of planetary bodies. This paper proposes a novel way of simultaneously utilizing optical images, digital elevation maps (DEMs), and slope maps for automatic crater detection on the lunar surface. Mask R-CNN, tuned for the crater detection task, is utilized in this paper. Two catalogs, namely, Head-LROC and Robbins, are used for the performance evaluation. Exhaustive analysis of the detection results on the lunar surface has been performed with respect to both Head-LROC and Robbins catalog. With the Head-LROC catalog, which has relatively strict crater markings and larger possibility of missing craters, recall value of 94.28\% has been obtained as compared to 88.03\% for the baseline method. However, with respect to a manually marked exhaustive crater catalog based on relatively liberal marking, significant precision and recall values are obtained for different crater size ranges. The generalization capability of the proposed method in terms of crater detection on a different terrain with different input data type is also evaluated. We show that the proposed model trained on the lunar surface with optical images, DEMs and corresponding slope maps can be used to detect craters on the Martian surface even with entirely different input data type, such as thermal IR images from the Martian surface.
In this paper, we present a domain adaptation based generative framework for Zero-Shot Learning. We explicitly target the problem of domain shift between the seen and unseen class distribution in Zero-Shot Learning (ZSL) and seek to minimize it by developing a generative model and training it via adversarial domain adaptation. Our approach is based on end-to-end learning of the class distributions of seen classes and unseen classes. To enable the model to learn the class distributions of unseen classes, we parameterize these class distributions in terms of the class attribute information (which is available for both seen and unseen classes). This provides a very simple way to learn the class distribution of any unseen class, given only its class attribute information, and no labeled training data. Training this model with adversarial domain adaptation provides robustness against the distribution mismatch between the data from seen and unseen classes. Through a comprehensive set of experiments, we show that our model yields superior accuracies as compared to various state-of-the-art ZSL models, on a variety of benchmark datasets.