[go: up one dir, main page]

Academia.eduAcademia.edu
PANSHARPENING OF MASTCAM IMAGES C. Kwan, B. Budavari, M. Dao, B. Ayhan J. F. Bell Applied Research LLC Arizona State University ABSTRACT This paper summarizes a new investigation of applying advanced pansharpening algorithms to enhance the images of the left imager in the Mastcam onboard the Curiosity rover, which landed on Mars in 2012. The various instruments on the rover have already made great contributions in the understanding of Mars. The goal of our research is to generate both high spatial and high spectral image cube by using the left and right Mastcam imagers. Eleven algorithms have been investigated using five objective performance metrics. Subjective evaluations have also been conducted. The image enhancement results are encouraging. Index Terms— Mastcam, pansharpening, image enhancement image registration, 1. INTRODUCTION The Curiosity rover has several instruments that can help characterize the Mars surface. The APXS [1] can analyze rock samples and extract compositions of rocks; the LIBS [2] can collect spectral features from the vaporized fumes and deduce the rock compositions at a distance of 7 m; and Mastcam imagers [3] can perform surface characterization from 1 km away. There are two Mastcam multispectral imagers, separated by 24.2 cm [3]. Specifically, the left Mastcam (34 mm focal length) has three times the field of view of the right Mastcam (100 mm focal length). That is, the right imager has 3 times higher resolution than that of the left. Each camera has 9 bands with 6 overlapping bands. For stereo image formation and image fusion (merging of the left and right bands to form a 12-band image cube), the current practice is to downsample the right images to the resolution of the left, avoiding artifacts due to Bayer pattern and also lossy JPEG compression. This practice is certainly practical, but may limit the full potential of Mastcams. First, although downsampling of the right images can preserve the spectral integrity and avoid certain artifacts of the image data, the process will throw away some high spatial resolution pixels in the right bands. Second, the current stereo images have lower resolution, which may degrade the augmented reality or virtual reality experience of science fans. If better 978-1-5090-4951-6/17/$31.00 ©2017 IEEE demosaicing/debayering and compression algorithms are available in the future, one may want to apply some advanced pansharpening algorithms to the left bands so that one can have 12 bands of high resolution image cube for stereo vision and image fusion. In recent years, significant advances have been made in image super-resolution and new and high performance pansharpening algorithms have been proposed regularly. In light of the above development, a natural research question is: can we keep the 9 high resolution bands in the right imager and improve the resolution of all the bands in the left imager while maintaining the spectral integrity of the left bands? If the answer is positive, this may motivate researchers to develop high performance demosaicing algorithms and also adopt more advanced compression algorithms such as X264 in future planetary missions. Moreover, both the user experience of using high resolution stereo images, as well as the surface characterization performance using 12-band of high resolution images, will be greatly enhanced. In this paper we attempt to answer the above question by applying some recent pansharpening algorithms to the Mastcam data. Preliminary results using actual Mastcam images are encouraging. Objective and subjective studies show that it is indeed possible to enhance the spatial resolution of left bands while maintaining the spectral integrity of the left bands. 2. MASTCAM Fig. 1: Normalized MSL/Mastcam system-level spectral response profiles for the left eye M-34 camera (top panel) and the right eye M-100 camera (bottom panel) [3]. 5117 IGARSS 2017 Mastcam imager information is shown in Fig. 1. There are 6 overlapping bands and 3 non-overlapping bands (L3, L4 and L5 from the left camera and R3, R4, and R5 from the right camera). More details can be found in [3]. that can be adapted to the Mastcam data and has great potential in significantly enhancing the resolution of left bands. The basic idea is to utilize the right bands to enhance the left bands. Low spatial resolution right bands 3. KEY ALGORITHMS 3.1. Image Registration Since the left image is of lower resolution, its resolution needs to be enhanced before its 3 non-overlapping bands can be merged together with the 9 high resolution bands in the right camera. Before we can perform image enhancement, the first step is to align the left and right image to subpixel accuracy; otherwise misalignment errors may propagate to the image enhancement process. A two-step image alignment approach was developed [4] and the signal flow is shown in Fig. 2. The first step of the two-step image alignment approach is using RANSAC (Random Sample Consensus) technique [19] for an initial image alignment. In this first step, we use the two RGB stereo images from the left and right Mastcams. First, SURF features [20] and SIFT features [21] are extracted from the two stereo images. These features are then matched within the image pair. This is followed by applying RANSAC to estimate the geometric transformation. Assuming the right camera image is the reference image, the left camera image content is then projected into a new image that is aligned with the reference image using the geometric transformation. The second step of the two-step alignment approach uses this aligned image with RANSAC and the left camera image as inputs and applies the Diffeomorphic Registration [4] technique. Diffeomorphic Registration is formulated as a constrained optimization problem, which is solved with a step-then-correct strategy [4]. This second step reduces the registration errors to subpixel levels so that pansharpening can be performed. High spatial resolution right bands LR R HR R LR L HR L Low spatial resolution left bands Pansharpened high spatial resolution left bands Fig. 3. System flow of color mapping. LR denotes low resolution; HR denotes high resolution; LR R denotes the set of low resolution right pixels; LR L denotes the set of low resolution left pixels; HR L denotes high resolution left pixels. Our idea [5] is called color mapping, which is the mapping of a right pixel c( i , j ) at location (i,j) with M bands to a left pixel X ( i , j ) with N bands at the same location. This mapping is based on a transformation matrix T, i.e. X ( i , j ) = T c( i , j ) (1) N ×M where T ∈ R . Fig. 3 shows the system flow. Given a high resolution (HR) right image cube and a low resolution (LR) left image cube, our goal is to generate a HR left image cube. One advantage of our algorithm is that the Bayering and JPEG artifacts will not be amplified, as we do not require bicubic interpolation of the left images in our algorithm. To get the transformation matrix, we simulate a low resolution right image cube by down-sampling the HR right image cube. We then use the LR right image cube and the LR left image cube to train the T. Once T is obtained, it can then be used for generating the HR left image pixel by pixel. Details of our algorithm can be found in [5][6][24]. Remark 1: HCM A variant of the color mapping is to introduce a white band and some low resolution left bands into the vector c( i , j ) . That is, ch = [c(1), …, c(M), h(k1), h(k2), ..., h(kt), 1]T (2) where [h(k1), h(k2),…, h(kt)] are extracted from the low resolution left image cube. kt is the number of selected bands. Fig. 2. Block diagram of the two-step image alignment approach. 3.2. Proposed Hybrid Color Mapping (HCM) A simple approach to increasing the resolution of the left images is to use bicubic interpolation [18]. However, it was shown that bicubic interpolation is coarse and does not yield satisfactory performance in both objective and subjective evaluations [5] [6]. We developed a new algorithm in [5] Remark 2: Local HCM We further enhance our method by applying color mapping patch by patch. A patch of size p × p is a sub-image in the original image. The patches do not overlap. In this way, spatial correlation can be exploited. As a result, the mapping will be even more accurate. In addition, since the task is split 5118 into many small tasks, the process can be easily parallelized and hence is suitable for real-time processing. 3.3. Other Pansharpening Algorithms There are two recent survey papers [7][8] summarizing the state-of-the-art pansharpening algorithms in the literature. The following 10 algorithms were selected in our studies: Smoothing Filter-based Intensity Modulation (SFIM) [9], Modulation Transfer Function Generalized Laplacian Pyramid (MTF-GLP) [10], MTF-GLP with High Pass Modulation (MTF-GLP-HPM) [11], Gram Schmidt (GS) [12], GS Adaptive (GSA) [13], Principal Component Analysis (PCA) [14], Guided Filter PCA (GFPCA) [15], Hysure [16], and partial replaced adaptive component substitution (PRACS) [17]. The bicubic interpolation method [18] is also included in our study. Although bicubic is faster than HCM, it is not considered as a pansharpening method, as it does not use any of the right bands. If we exclude bicubic, then HCM yielded the best performance in all categories. Fig. 4 shows the CC of different methods. CC is a measure of spectral integrity of pansharpening. It can be seen that our HCM method performs well across almost all the bands. Table 1: Comparison of HCM with other algorithms. HCM SFIM MTF-GLP MTF-GLP-HPM GS GSA PCA GFPCA Hysure PRACS Bicubic 3.4. Performance Metrics We used the following 5 performance metrics in our studies: Root Mean Squared Error (RMSE) [7], Cross Correlation (CC) [7], Spectral Angle Mapper (SAM) [7], Erreur relative globale adimensionnelle de synthèse (ERGAS) [7], and computational time. Moreover, subjective evaluation has been carried out for those non-overlapping left bands that do not have ground truth images. 4. COMPARATIVE STUDIES 4.1. Data The Mastcam dataset downloaded from the Planetary Data System (PDS) resource contains a total of more than 500,000 images collected at different times and locations. Since left and right Mastcams are independently controlled and do not always collect data simultaneously, we have to perform extensive pre-processing, which exhaustively screens through all images to only select pairs of image sets that consist of images of all available spectral bands in both left and right Mastcam cameras. After preprocessing and cleaning up the image sets, we can construct a total of 133 LR-pairs. 4.2. Performance of Pansharpening with Known Ground Truth For illustration purpose, we show the full details of the pansharpening results of using one pair of Mastcam images collected on sol 150 in our comparative studies. In HCM, a patch size of 151 was used and only the 6 overlapping bands in the right Mastcam were used. In addition, we performed two iterations of the HCM algorithm to further improve its performance. Similarly, for all the other pansharpening algorithms, the pan band was created by taking the mean of the 6 overlapping right bands. Table 1 shows the 5 metrics. The RMSE, CC, SAM, and ERGAS were generated by using the right high resolution bands as the ground truth. It can be seen that HCM performed the best in all categories, except time. RMSE 0.0287 0.0368 0.0351 0.035 0.0333 0.0316 0.0311 0.0346 0.0312 0.0336 0.0406 CC 0.9693 0.9192 0.9308 0.9314 0.9403 0.9521 0.9529 0.9306 0.9563 0.9409 0.8902 SAM 2.3201 2.7123 2.6969 2.6931 2.631 2.6197 2.6242 2.5592 2.6617 2.7438 2.8083 ERGAS 1.4723 1.9276 1.8221 1.8153 1.7294 1.6075 1.6072 1.809 1.5971 1.7281 2.1489 Time 1.0661 1.7446 1.2508 1.3785 1.1254 1.2033 1.2436 1.4508 483.08 6.6205 0.0688 Fig. 4: Comparison of CC between HCM and all methods. 4.3 Performance of Image Fusion without Reference The 3 non-overlapping bands in the left imager do not have ground truth data. In [8], a metric known as Quality with No Reference (QNR) was described. However, QNR requires the pan to be overlapped with the other bands. Recently, a new blind image quality assessment tool has been developed [23]. We plan to apply that to this application in the future. Here, we performed subjective evaluations by displaying the pansharpened images. Fig. 5 shows the false color images by forming RGB images using the pansharpened images from L5, L3, and L4 bands. It can be seen that Hysure and HCM can give very good visual appearance. If one zooms in, one will notice that there are color distortions in GSA, GS, and other methods whereas Hysure and HCM show almost no observable color distortion. 5. CONCLUSIONS The application of recently developed pansharpening algorithms to enhance the left Mastcam images has been summarized in this paper. Using both objective and 5119 subjective performance metrics, preliminary results showed that image quality can be improved tremendously. One future direction is to develop a metric that can assess the pansharpening performance without reference. Another direction is to assess the performance gain by using the fused image cube formed by combining the HR right bands with the enhanced left bands in some pixel clustering and anomaly detection applications [22]. A third one is to utilize point spread function (PSF) to further enhance the resolution of the left images. HCM SFIM MTF-GLP GS GSA PCA Hysure PRACS bicubic MTF-GLP-HPM GFPCA Fig. 5. Subjective evaluation of different pansharpening algorithms. False color images are shown here. 6. REFERENCES [1] B. Ayhan, C. Kwan, and S. Vance, “On the Use of a Linear Spectral Unmixing Technique for Concentration Estimation of APXS Spectrum,” J. Multidisciplinary Engineering Science and Technology, Volume 2, Issue 9, September, 2015. [2] W. Wang, S. Li, H. Qi, B. Ayhan, C. Kwan, and S. Vance, “Revisiting the Preprocessing Procedures for Elemental Concentration Estimation based on CHEMCAM LIBS on MARS Rover,” 6th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Lausanne, Switzerland, June 24-27, 2014. [3] J. Bell et al., “The Mars Science Laboratory Curiosity Rover Mast Camera (Mastcam) Instruments: Pre-Flight and In-Flight Calibration, Validation, and Data Archiving,” submitted to the AGU journal Earth and Space Science, 23 September 2016. [4] B. Ayhan, M. Dao, C. Kwan, H. Chen, J. F. Bell, and R. Kidd, “A Novel Image Alignment Approach for Mastcam Imagers of Mars Rover with Application to Rover Guidance,” submitted to IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, April, 2017. [5] J. Zhou, C. Kwan, and B. Budavari, “Hyperspectral Image Super-Resolution: A Hybrid Color Mapping Approach,” SPIE Journal of Applied Remote Sensing, September, 2016. [6] C. Kwan, J. H. Choi, S. Chan, J. Zhou, and B. Budavari, “Resolution Enhancement for Hyperspectral Images: A SuperResolution and Fusion Approach,” Int. Conf. Acoustics, Speech, and Signal Processing, March, 2017. [7] L. Loncan, et al., “Hyperspectral Pansharpening: A Review,” IEEE Geoscience and remote sensing magazine, 9/2015. [8] G. Vivone, et al., “A Critical Comparison Among Pansharpening Algorithms,” IEEE Trans. Geoscience and Remote Sensing, Vol. 53, No. 5, May 2015. [9] J. G. Liu, “Smoothing filter based intensity modulation: A spectral preserve image fusion technique for improving spatial details,” Int. J. Remote Sens., vol. 21, no. 18, Dec. 2000. [10] B. Aiazzi, et al., “MTF-tailored multiscale fusion of highresolution MS and pan imagery,” Photogramm. Eng. Remote Sens., vol. 72, no. 5, pp. 591–596, May 2006. [11] G. Vivone, et al., “Contrast and error-based fusion schemes for multispectral image pansharpening,” IEEE Trans. Geosci. Remote Sensing Lett., vol. 11, no. 5, pp. 930–934, May 2014. [12] C. Laben and B. Brower, “Process for enhancing the spatial resolution of multispectral imagery using pan-sharpening,” U.S. Patent 6 011 875, Jan. 4, 2000. [13] B. Aiazzi, et al., “Improving component substitution pansharpening through multivariate regression of MS+pan data,” IEEE Trans. Geosci. Remote Sensing, vol. 45, no. 10, pp. 3230–3239, Oct. 2007. [14] P. S. Chavez, et al., “Comparison of three different methods to merge multiresolution and multispectral data: Landsat TM and SPOT panchromatic,” Photogramm. Eng. Remote Sens., vol. 57, no. 3, pp. 295–303, Mar. 1991. [15] W. Liao, et al., “Processing of multiresolution thermal hyperspectral and digital color data: Outcome of the 2014 IEEE GRSS data fusion contest,” IEEE J. Select. Top. Appl. Earth Observ. Remote Sensing, vol. 8, no. 6, May 2015. [16] M. Simões, et al., “A convex formulation for hyperspectral image superresolution via subspace-based regularization,” IEEE Trans. Geosci. Remote Sensing, 2015. [17] J. Choi, et al., “A new adaptive component-substitution based satellite image fusion by using partial replacement,” IEEE Trans. Geosci. Remote Sens., vol. 49, no. 1, Jan. 2011. [18] R. Keys, “Cubic convolution interpolation for digital image processing,” IEEE Transactions on Acoustics, Speech, and Signal Processing, vol. 29, no. 6, pp. 1153–1160, Dec. 1981. [19] R. Hartley and A. Zisserman, Multiple View Geometry in Computer Vision, Cambridge University Press, 2003. [20] H. Bay, et al., “SURF: Speeded Up Robust Features,” Computer Vision and Image Understanding (CVIU).Vol. 110, No. 3, pp. 346–359, 2008. [21] D. G. Lowe, “Object recognition from local scale-invariant features.” IEEE International Conference on Computer Vision, vol. 2, pp. 1150-1157, 1999. [22] J. Zhou, C. Kwan, B. Ayhan, and M. Eismann, “A Novel Cluster Kernel RX Algorithm for Anomaly and Change Detection Using Hyperspectral Images,” IEEE Trans. Geoscience and Remote Sensing, Vol.54, Nov. 2016. [23] C. Kwan, B. Budavari, A. Bovik, and G. Marchisio, “On Pansharpening and Blind Quality Assessment of Pansharpened WorldView-3 Images,” submitted to IEEE Geoscience and Remote Sensing Letters, May, 2017. [24] C. Kwan, B. Budavari, and G. Feng, “An Alternative Approach to Fusing MODIS and Landsat Images for Forward Prediction,” submitted to IEEE Geoscience and Remote Sensing Letters, March 2017. 5120