Data system design for a hyperspectral imaging mission concept

Share Embed


Descrição do Produto

See discussions, stats, and author profiles for this publication at: https://www.researchgate.net/publication/224407798

Data system design for a hyperspectral imaging mission concept CONFERENCE PAPER · APRIL 2009 DOI: 10.1109/AERO.2009.4839507 · Source: IEEE Xplore

CITATIONS

READS

2

43

8 AUTHORS, INCLUDING: Jennifer Carpena-Núñez

Christianna Taylor

University of Puerto Rico at Rio Piedras

Skolkovo Institute of Science and Technology

12 PUBLICATIONS 56 CITATIONS

11 PUBLICATIONS 5 CITATIONS

SEE PROFILE

SEE PROFILE

Charles D. Norton California Institute of Technology 67 PUBLICATIONS 453 CITATIONS SEE PROFILE

All in-text references underlined in blue are linked to publications on ResearchGate, letting you access and read them immediately.

Available from: Jennifer Carpena-Núñez Retrieved on: 03 February 2016

Data system design for a hyperspectral imaging mission concept

Citation

Hartzell, C.M. et al. “Data system design for a hyperspectral imaging mission concept.” Aerospace conference, 2009 IEEE. 2009. 1-21. © 2009 Institute of Electrical and Electronics Engineers.

As Published

http://dx.doi.org/10.1109/AERO.2009.4839507

Publisher

Institute of Electrical and Electronics Engineers

Version

Final published version

Accessed

Fri Apr 20 05:36:48 EDT 2012

Citable Link

http://hdl.handle.net/1721.1/59373

Terms of Use

Article is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.

Detailed Terms

Data System Design for a Hyperspectral Imaging Mission Concept . Christine M. Hartzell Aerospace Engineering Sciences University of Colorado Boulder, CO 80309 (206) 915-6409 [email protected]

Jennifer Carpena-Nunez Department of Physics University of Puerto Rico San Juan, PR 00931 (787) 381-4171 [email protected]

Lindley C. Graham Department of Aeronautics and Astronautics Massachusetts Institute of Technology Cambridge, MA 02139 (408) 306-9247 [email protected]

David M. Racek Electrical and Computer Engineering Department Montana State University Bozeman, MT 59717 (253) 205-4538 [email protected]

Tony S. Tao Aerospace Engineering Pennsylvania State University University Park, PA 16802 (610) 306-2761 [email protected]

Christianna E. Taylor School of Aerospace Engineering Georgia Institute of Technology Atlanta, GA 30332 (617) 669-6774 [email protected]

Hannah R. Goldberg Jet Propulsion Lab California Institute of Technology Pasadena, CA 91109 (818) 216-6856 [email protected]

Charles D. Norton Jet Propulsion Lab California Institute of Technology Pasadena, CA 91109 (818) 793-3775 [email protected]

complexity of Earth science increases. The designs presented here are the work of the authors and may differ from the current mission baseline.

Abstract—Global ecosystem observations are important for Earth-system studies. The National Research Council’s report entitled Earth Science and Applications from Space is currently guiding NASA’s Earth science missions. It calls for a global land and coastal area mapping mission. The mission, scheduled to launch in the 2013-2016 timeframe, includes a hyperspectral imager and a multi-spectral thermal-infrared sensor. These instruments will enable scientists to characterize global species composition and monitor the response of ecosystems to disturbance events such as drought, flooding, and volcanic events. Due to the nature and resolution of the sensors, these two instruments produce approximately 645 GB of raw data each day, thus pushing the limits of conventional data handling and telecommunications capabilities. The implications of and solutions to the challenge of high downlink data volume were examined. Low risk and high science return were key design values. The advantages of onboard processing and advanced telecommunications methods were evaluated. This paper will present an end-to-end data handling system design that will handle the large data downlink volumes that are becoming increasingly prevalent as the

TABLE OF C ONTENTS 1 2 3 4 5 6 7 8

M ISSION OVERVIEW . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . C LOUD D ETECTION T ECHNIQUES . . . . . . . . . . . . . . . C OMPRESSION S CHEMES . . . . . . . . . . . . . . . . . . . . . . . . . C OMMUNICATIONS A RCHITECTURE . . . . . . . . . . . . . G ROUND S TATION S ELECTION . . . . . . . . . . . . . . . . . . . S CIENCE A NALYSIS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . DATA S TORAGE AND D ISTRIBUTION . . . . . . . . . . . . . C ONCLUSION . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ACKNOWLEDGEMENTS . . . . . . . . . . . . . . . . . . . . . . . . . . . R EFERENCES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

1 4 5 8 12 14 16 18 19 19

1. M ISSION OVERVIEW The National Research Council’s (NRC) report entitled Earth Science and Applications from Space provides a roadmap of Earth-science missions for the next decade. The HyspIRI

c 978-1-4244-2622-5/09/$25.00 2009 IEEE. IEEEAC Paper #1375, Version 4 Updated 1/09/2009.

1

mission, expected to launch between 2013 and 2016, is called for by the NRC to monitor global ecosystem health in order to identify the ecological effects of climate change and urbanization. The nominal mission duration of 3 years will allow analysis of the temporal variation in the ecological properties as well. The science objectives are achieved using two imaging spectrometers that operate in different regions of the electromagnetic spectrum. Due to the high resolutions of the imagers and the global mapping nature of the mission, approximately 6455 gigabytes of raw data will be collected each day.

The HyspIRI payload is sized to be approximately 145 kg. A commercial bus, the SA-200HP bus from General Dynamics, will be modified for use on the HyspIRI mission. Notable characteristics of this bus include upgrade options for GPS and the ability to have 3-axis pointing with 16 arcsec pointing control, 0.5 arcsec knowledge, and 0.1 arcsec/sec stability. The bus should provide 2,000 W at end-of-life (EOL) (at 1 AU) and has enough onboard fuel capacity for a 6-year mission. The launch vehicle is assumed to be a Taurus-class rocket [1]. Based on the last TeamX study conducted in July 2007, the HyspIRI satellite will orbit at an altitude of 623 km in a circular, sun-synchronous orbit, which translates into an inclination of 97.0 degrees [1]. The local descending node of the orbit would be 11am [1]. According to the mission scientists, with this local descending node time, there will be the fewest clouds (statistically) obstructing the sensors’ view of the Earth [2].

Science Objectives HyspIRI aims to detect ecosystem responses to climate change and human land management. It intends to monitor and enable prediction of volcanic activity, landslide hazards, wildfire susceptibility, drought, exploitation of natural resources, species invasion, surface and soil alterations, changes in surface temperature, deforestation and ecosystem disturbance. These can be detected by distinctive spectral features in the shortwave, visible, near infrared and infrared regions of the electromagnetic spectrum.

Missions Operations A selective instrument resolution scheme was created to reduce the mission’s data volume while retaining image detail in more scientifically-significant areas. By working with the mission scientists, Robert Green and Simon Hook at JPL, geographic areas where the data resolution could be reduced were identified.

Imager Characteristics The HyspIRI mission carries two imaging spectrometers: the Visible Shortwave Infra Red (VSWIR) hyperspectral imaging spectrometer and a Thermal Infra Red (TIR) multispectral imager. The two instruments will address independent science objectives.

Since the TIR measures the thermal energy radiated by the earth, it will operate during the entire orbit, regardless of the position of the sun. If the sensor is taking measurements of the ocean, the resolution will be reduced by averaging the data over multiple pixels. The spatial resolution over the ice sheet in Antarctica is also reduced to the ‘Ocean’ resolution level. According to Hook, 200 m x 200 m would be a desirable resolution for the TIR over the ocean [3]. Leastsignificant spectral quantization data will be dropped to reduce the quantization resolution to two bits per band since exact spectral data is not as important to scientists as an overall classification and trends in the regions of lower resolution. Cloud data is of interest to scientists in this spectral range, so clouds will not undergo data reduction [3]. The TIR’s resolution modes are enumerated in Table 1 below. Using this scheme, 97% of TIR data volume would be devoted to Land/Shallow Water data.

The VSWIR imager will capture sunlight reflected off the Earth’s surface, while the TIR collects thermal energy emitted by the Earth’s surface. The VSWIR instrument will have a 145 km wide swath width. It is a nadir-pointing push-broom instrument, with a spatial resolution of 60m and a repeat cycle of 19 days. The TIR instrument will have a 600 km wide swath width. It is also nadir-pointing, however it has a shorter repeat cycle (5 days) due to its large swath width. The TIR instrument is a whisk-broom sensor with a spatial resolution of 60m. The VSWIR instrument will collect reflected light at 214 spectral bands, from 380nm (violet) to 2510nm (near infrared) with a 10nm spectral resolution. The TIR collects data at 8 bands, 7 bands between 7.3-12.1 µm and 1 band between 3-5 µm. Due to the locations of their spectral bands in the electromagnetic spectrum, the TIR will address science objectives concerning thermal events while the VSWIR will address science objectives related to surface composition and ecosystem information.

The VSWIR instrument’s resolution modes differ from those of the TIR instrument. Since the VSWIR instrument requires the reflected light from the sun to operate, the sensor will only be active when the sun-elevation angle is sufficient to produce meaningful data [4]. Current sensor design puts this elevation at 20◦ . Resolution modes for the VSWIR are shown below in Table 2. Clouds imaged by the VSWIR instrument are compressed spectrally since images of clouds in the VSWIR wavelengths provides little scientifically usefully information. Cloud pixels are not compressed spatially since the spatial unpredictability of clouds would make consistent

Mission Design Characteristics The preliminary mission design presented here was developed by JPL’s TeamX. TeamX creates integrated preliminary mission design concepts using a collaborative engineering environment. 2

Table 1. Spatial, spectral, and quantization resolutions for the TIR instrument by land classification.

Table 2. Spatial, spectral, and quantization resolutions for the VSWIR instrument by land classification.

spatial compression impossible. Using this scheme, 98% of VSWIR data volume would be devoted to Land/Shallow Water data. Note that, in contrast to the TIR, the VSWIR keeps its full spectral and quantization resolution over ‘Ocean’ – designated areas.

to store hundreds of gigabytes of information until the next ground station is in range. The communications system will have to handle transmission of five terabits of data per day, yet still have a robust enough downlink margin to allow for a missed communications pass. Once the data is on the ground, a series of ground calibrations and processing must be done and the challenge again becomes the huge data volume to be stored. The problem of making the science products readily available and accessible is not easily answered when sifting through hundreds of terabytes of data.

Data Volume Challenge Hyperspectral data is not new to earth science remote sensing missions, rather what makes the HyspIRI mission unique is the amount of data being collected, stored, and distributed. The huge data rate, on the order of hundreds of gigabytes per day, is due to the two instruments collecting high resolution data over a wide spectral range. The spectral information, in the case of the VSWIR imager, is quantized into 214 bands each represented by a 14 bit sample. The spectral information collected by the TIR imager is similarly quantized into 8 bands, with each sample being described in 14 bits. With a maximum sample rate of approximately 70Msamples/second, the resulting data stream is substantial.

The HyspIRI mission will collect nearly two petabytes over the three year mission life. Figure 1 shows the overall data reduction transferred over the communication system and the data stored on the ground over the three year mission. The data reduction that is accounted for in these graphs includes reductions due to decreasing the instrument resolution when taking data over areas of little scientific interest and onboard compression of data. Shown on the left is the amount of data reduced on-board the spacecraft per orbit, thereby reducing the volume of data that must be transmitted to the ground. On the right is the total data reduction over the three year mission. It can be seen that there is a substantial reduction in data volume due to the compression techniques employed and discussed in detail in this paper.

The HyspIRI mission presents a data handling challenge that will push the limits of current processing, storage, and communication technologies. Moore’s law, stating that the number of transistors that can be inexpensively placed on an integrated circuit will double approximately every two years [5], may ease the cost and technology boundaries by the mission’s 2011 technology cutoff. Nevertheless, the downlink capabilities and on-board processing power of the HyspIRI mission limit the amount of data that can be collected and analyzed. The mission utilizes hardware implementation of data processing and compression methods to reduce the stream of data coming from the instruments. The on-orbit data system must be able to process and compress up to 70M samples per second in real time. Different resolution modes, cloud screening, and lossless data compression become vital to the overall data reduction. Without these data reduction methods, the data volume would be well beyond the feasible volume to transmit to earth, meaning a smart combination of these options must be designed and modeled. The spacecraft is physically limited by the on-board solid state storage which must be able

Figure 1. Data Reduction Graphs showing the effectiveness of the compression schemes employed by the HyspIRI mission. Given the high data downlink volume, a two-pronged ap3

proach was taken in designing the mission’s data system. First, techniques were employed to reduce the data downlink volume. Once those techniques were exhausted, a data system architecture was designed to accommodate the resulting data downlink volume. Five topics associated with data downlink volume reduction and accommodation were investigated in further detail and presented here. In attempts to reduce the data downlink volume, cloud detection and hyperspectral data compression algorithms were identified. In order to accommodate the resulting data downlink volume, communications architectures were analyzed and ground stations were identified. Additionally, preliminary science products were identified and a ground-based data storage architecture was developed.

scattering effects. Further analysis, possibly in the form of an Observing Simulated Systems Experiment (OSSE), could investigate the effect of this omission on the quality of the science returns.

2. C LOUD D ETECTION T ECHNIQUES An on-board cloud detection algorithm is required for the HyspIRI mission in order to reduce the data downlink volume. While the cloudy TIR data has science value, there is little benefit to returning cloudy data in the spectral range of the VSWIR instrument. Since clouds cover nearly 30% of the globe, the total mission data downlink volume can be reduced by nearly 96 GB (13%) per day by applying additional compression to the cloudy pixels. In order to detect clouds, the raw data must first be converted to the top-of-atmosphere (TOA) reflectance. There are several detection algorithms that can be compared and evaluated based on the needs of the mission. Finally, once clouds have been detected, the cloudy pixels will undergo lossy spectral and bitwise compression, before being further compressed by the main compression algorithm. Note that clouds will only be detected when the instrument is over land or coastal areas, due to the selective resolution of the instrument. A preliminary cloud detection algorithm was applied to sample hyperspectral data collected by an air-borne imaging spectrometer (JPL’s Airborne Visible/Infrared Imaging Spectrometer (AVIRIS)) as a proof-ofconcept trial.

Figure 2. Calculation of Reflectance for the Equivalent Solar Zenith Angle (θz ). Reflectance is the sensed radiance divided by the normal component of the solar radiance for the zenith angle.

Types of Cloud Detection Algorithms The cloud detection algorithm selection was driven by two main requirements: the algorithm must be able to process data in real-time on-board the spacecraft and it must distinguish between cloud and non-cloud pixels with sufficient accuracy. Specifically, since the cloudy pixels will be lossily compressed, the cloud detection algorithm must be clear-sky conservative; it is more important to avoid erroneously classifying pixels as cloudy than to detect all the cloudy pixels.

Preliminary On-board Processing

On first glance, the problem of cloud detection seems simple, given the characteristic shapes and brightness of clouds. However, automated cloud detection poses a challenge since classifiers must be able to discern on a pixel-by-pixel basis cloud from non-cloud. With respect to the high reflectance of clouds, in hyperspectral imagers it is a challenge to determine at which wavelength to look for a characteristic cloud signature since smoke, ash plumes, snow and highly-reflective sand can be easily mistaken for clouds. Often cloud detection algorithms are applied once the data has been sent to the ground, where there are fewer constraints on processing time and memory requirements.

In order to apply a cloud detection algorithm, the raw VSWIR data must be converted to TOA reflectance. First, the dark level signal will be subtracted from the raw data. Then the data is converted to radiance. Finally, the radiance is converted to TOA reflectance for the equivalent solar zenith angle by dividing the sensed radiance by the normal component of the solar radiance for the equivalent zenith angle (see Figure 2). Similar on-board calibration was done during the Hyperion instrument’s extended mission on the Earth Observing-1 satellite [6]. The equivalent solar zenith angle is calculated on-board using the spacecraft’s ephemeris data. Although, it is theoretically necessary to recalculate the zenith angle for each pixel, it is predicted that impact of recalculating this value less frequently will be minimal. Further analysis should address the implementation of this calibration process onboard the spacecraft. Also, it is important to note that this calibration protocol does not take into account the atmospheric

Cloud detection is a subset of current research surrounding automated pixel classification. There are several types of algorithms that can be used for pixel classification. A comparison of four types of pixel classification, used for the purpose of autonomous event identification, is given in [7]. A more detailed look at cloud detection algorithms is given in [8]. 4

On-board cloud detection was previously implemented on the Hyperion instrument of Earth Observing-1 (EO-1). An algorithm containing a series of thresholds was initially implemented on the instrument (see [9] for details) for cloud detection. During Hyperion’s extended mission, a Support Vector Machine algorithm was used for pixel identification [6]. Due to the heritage of these two algorithm types on-board Hyperion, they were selected for further investigation with respect to implementation on the HyspIRI mission.

detection algorithms that could be used by HyspIRI. A simple threshold algorithm was applied to the sample AVIRIS scenes: if ρ1.67µm ≥ 0.5, then the pixel is cloudy, where ρx is the reflectance at wavelength x. The threshold was applied at a wavelength (1.67µm) where clouds have high reflectance values and there is a low sensitivity to the presence of water. The algorithm was first applied to an AVIRIS image from South-Central Pennsylvania (39.96◦ N, -78.32◦ E) taken July 2, 2008. This image is an obviously cloudy, agricultural scene. The accuracy of the cloud mapping algorithm can be seen in Figure 3. It can be seen that most of the clouds are detected by the algorithm. Figure 3b, a subset of Figure 3a, shows that the algorithm does not very accurately detect smaller, less bright clouds. Unfortunately, no quantitative analysis can be done on the results from this classification because there is currently no “truth” data - i.e., the image has not been manually labeled for clouds. However, the preliminary qualitative results from this image indicate that the classifier is performing as desired, that is to say, it is clear-sky conservative.

Threshold cloud detection algorithms operate like a series of logic gates. Pixels are labeled as clouds if their reflectances are within a given range for a given wavelength or set of wavelengths. At its simplest, a threshold algorithm can take the form: if ρ1.67 ≥ 0.4, then the pixel is cloudy. A series of six thresholds was initially implemented on Hyperion for cloud detection [9]. Multiple thresholds were used in order to reduce the classification error over desert, snow, ice and highly reflective water. The advantage of threshold algorithms is that their simplicity results in fast processing times. However, threshold algorithms become impractical if attempting to label more than one class of pixels. Support Vector Machines (SVMs) are more capable classifiers than threshold algorithms, although this increased capability comes with a cost of increased complexity. SVMs can be used to label multiple classes of pixels, for example snow, water, ice, land and cloud. An SVM is a machine learning algorithm that maps data into a higher dimensional space and then classifies it based on individual data points’ positions relative to separating hyperplanes [10]. The SVM must be trained using sample data to distinguish between the different classes of pixels. This can be accomplished through user-defined pixel labeling or clustering methods, which classify the pixels based on the desired number of classifications. SVMs also vary in complexity based on the way they separate the data in multidimensional space. A more detailed explanation of the theory behind SVMs can be found in [11]. On EO-1, an SVM using a linear kernel was implemented to autonomously detect science events of interest, such as volcanic eruptions and ice sheet breakup. Similar capabilities could be implemented on HyspIRI, although they are not specifically required by the current science objectives.

The algorithm was also applied to a desert scene from Central Nevada (38.52◦ N, -117◦ E) taken on June 11, 2008. As mentioned, sand, which can be highly reflective, is often difficult to differentiate from cloudy pixels. For this scene, which contained no clouds, the classifier erroneously labeled 0.11% of the pixels as cloudy. The cloud detection must be applied to more data samples in order to determine the average classification error. Additionally, the minimum classification accuracy that is required should be defined by the science community. The required algorithm accuracy may vary based on the ground cover due to the mission’s science objectives. For instance, incorrectly classifying pixels as clouds over the desert will have a limited impact on the global plant species map science product. The HyspIRI mission will use a threshold algorithm to detect clouds in data collected by the VSWIR instrument. Preliminary work presented here has shown that threshold algorithms have suitable accuracy in detecting clouds. Once the cloudy pixels have been detected, they will be lossily compressed by reducing the spectral and radiometric resolutions of the data.

A threshold cloud detection algorithm has been baselined for the HyspIRI mission due to its simplicity. Since the classifier is only required to classify pixels into two classes (cloud and non-cloud), a threshold algorithm can meet the mission’s needs.

3. C OMPRESSION S CHEMES Given the large volume of data to be produced by HyspIRI, on-board data compression is critical to enabling the mission. When both instruments are operating in full resolution mode, the HyspIRI onboard processing system must be able to process up to 70Msamples/second in real time before the data can be sent to the solid state recorder and prepared for downlink by the communications system. Based on the last JPL Team X design, onboard data storage is limited to 1.6 Terrabits, or 200 Gigabytes, of solid state memory [1]. This memory constraint, along with the maximum downlink rate capa-

Preliminary Algorithm Results In order to demonstrate the feasibility of the proposed preliminary calibration and threshold cloud detection algorithm, the process was applied to two scenes of AVIRIS data. Although AVIRIS data does differ slightly from the data that would be collect by the VSWIR instrument, it is similar enough in terms of wavelengths of spectral bands to be used to test the 5

Figure 3. Cloud Detection Results for Pennsylvania AVIRIS Data. a) shows the full image swath, with true color on left with the cloud detection algorithm results on the right. b) is a close-up from same data set. It can be seen that the algorithm has some difficulty detecting small clouds and cloud edges. bilities are the two main drivers of the need for onboard real time data compression for the HyspIRI mission. The specific requirements on the compression algorithm as well as a comparison between select algorithms will be presented.

because they are of less scientific interest). Since the instrument is designed to collect high resolution data, the compression should not degrade that resolution. 2. Does the algorithm function fast enough to be implemented in real time with a high throughput data rate?

Compression Algorithm Requirements The compression algorithm was chosen in response to three driving requirements:

The speed requirement of the algorithm is a simple result of the data rate at which the sensors will collect data. An algorithm that is fast will have low computational requirements. A fast algorithm will also utilize hardware implementation, meaning all of its operations can be easily described in a hardware description language and implemented on an integrated circuit or field programmable gate array (FPGA).

1. Does the algorithm effectively compress hyperspectral data in a lossless manner? As expected, it is desirable to implement the highest degree of compression possible without losing the scientific usefulness of the data. Lossless compression allows a reduction of data volume while still maintaining the ability to exactly reconstruct the original file. In this application lossless compression is necessary to preserve the scientific integrity of the data collected. Among the things that set this mission apart from previous remote sensing missions is the fine spatial resolution of the instruments. Therefore, in order to produce a superior end-level data product it is important that at no point is the data compressed in a lossy manner (Note that this approach is for the land and shallow water regions only. Cloudy areas taken by the VSWIR instrument are lossily compressed

3. Can the algorithm be implemented on-board? Since the algorithm will be implemented on-board the spacecraft, it must have low computational and memory requirements. Each sample will require some number of computational operations when being compressed; this number should be on the order of tens of operations per sample [12]. In addition, multiplication operations take a larger number of clock cycles so an efficient algorithm will be optimized to use few multiplication operations and more add, subtract, and bit shift 6

operations.

cloud map is then sent to the solid state recorder. If the compressor is compressing a cloudy pixel, the pixel is initially lossily compressed, then additional lossless compression is applied. The initial compression works by first averaging the 214 spectral bands to 20 and reducing each sample from 14 to 4 bits. If the compressor is compressing a non cloud pixel, only the lossless compression algorithm is applied. The compressed data stream is sent to the solid state recorder, where it will be stored until it can be transmitted to a ground station. The compression algorithm will be implemented on a field programmable gate array (FPGA).

Fast Lossless Compression Algorithm Five lossless compression algorithms were compared in terms of compression ratio by Aaron Kiely at JPL (see Table 3). The data compression algorithms were tested using calibrated 1997 16 bit AVIRIS data from the Jasper Ridge, Lunar Lake and Moffett Field datasets. The table shows the average number of bits/sample to which the data was reduced. The Fast Lossless algorithm, developed by Matt Klimesh at JPL, had the lowest number of output bits per sample and the overall best performance. The average compression ratio achieved by the Fast Lossless algorithm for these datasets was 3.1:1. Traditional algorithms that have previously been used on hyperspectral data produced a higher average bits/sample and less effective compression ratios [13]. For detailed information on the Fast Lossless Compression algorithm, please see [12].

Figure 4. Data Flow through the compressor for the VSWIR instrument.

Hardware Implementation The motivation for hardware implementation of the onboard compression algorithm is speed. Hardware compression may perform as much as five times faster than software compression for data of this type [12]. To implement the algorithm in hardware, it is translated from C code (executed as a list of software instructions) to a hardware description language (executed as a propagation of signals through circuits). The low complexity and minimal computational requirements of the Fast Lossless compression algorithm make it very well suited to be implemented using an FPGA.

Table 3. Comparison in terms of resulting bits per sample after processing by compression algorithms. Initially, all samples were initially expressed with 16 bits.

In addition to supplying an advantage in terms of compression ratio over other algorithms, the Fast Lossless Compression Algorithm is also well suited for implementation on an FPGA due to the algorithm’s low number of calculations per sample. The memory requirements are also low, requiring only enough memory to hold one spatial-spectral slice of the data (on the order of hundreds of Kbytes) [12].

Work to implement the Fast Lossless Compression Algorithm on an FPGA is being done at JPL. The algorithm was tested using a Xilinx Virtex 4 LX160 FPGA mounted on a custom made evaluation board and preliminary test results demonstrated a throughput data rate of 33 Msamples/sec at a system clock rate of 33MHz. The algorithm used approximately ten percent of the chips’ available programmable logic and interconnect resources [12]. The VSIWIR instrument will collect samples at a maximum rate of 60 Msamples/sec, the TIR will collect at a maximum of 10 Msamples/sec, for a combined maximum sample rate of 70 Msamples/sec, which should be within the capabilities of the board.

Figure 4 shows the data flow from the instrument to the solid state recorder in the case of the VSWIR instrument. Data from the sensor is initially calibrated to reflectance values which can be analyzed by the cloud detection algorithm. Calibration data is sent to the solid state recorder. The cloud detection algorithm generates a cloud map that becomes an input to the compressor. The cloud map can be thought of as a matrix of cells that correspond to a spatial location on the image. Each cell holds one bit to tell the compressor if the current pixel is a cloud pixel or a non cloud pixel. The

FPGAs also provide appealing features such as high density, low cost, radiation tolerance, and functional evolution [14]. As science goals change or as spacecraft capabilities are limited, the ability of an FPGA to be reprogrammed from earth allows for the functional evolution of hardware thought the 7

life of the mission. For the HyspIRI mission this means that if a new compression algorithm with a higher compression ratio or a more accurate cloud detection algorithm with better cloud classification is developed, the new technology could be implemented on the satellite with relative ease. In March 2008, Xilinx Inc., the main manufacturer of space ready FPGAs, has released the Radiation-Tolerant Virtex-4 QPro-V Family, providing good options for radiation-tolerant FPGAs with the speed and processing power required for the HyspIRI mission [15]. Thus, based on a comparison of available compression algorithms and the HyspIRI missions requirements, an FPGA implementation of the Fast Lossless Compression Algorithm is the recommended compression scheme for the mission.

that clouds will only be identified on the VSWIR instrument when it is viewing LSW regions. Some key elements of the model are shown in Table 4.

4. C OMMUNICATIONS A RCHITECTURE Once the data has been compressed, it will be transmitted to a series of ground stations using a given communications protocol. X-band, Ka band, TDRSS, and LaserComm architectures were compared in terms of their ability to accommodate the volume of data to be transmitted, their cost and mass, and their heritage level. An X-band architecture utilizing three ground stations was chosen. Data Flow Modeling Table 4. Important values and assumptions in the Data Flow Model.

In order to design the communications system, it was necessary to determine the desired data volume per unit time that must be transmitted from the satellite to the ground. To do this, a data flow model was created to calculate the data rate and volume that would be created aboard the spacecraft as seen in Figure 5.

The bits per orbit for each of the instrument and modes were then summed over one day. Note that the VSWIR is considered to be ‘on’ for half the orbit; the actual duty cycle would be less due to the minimum 20◦ sun angle. Therefore, the model will slightly overestimate the data taken over the ocean. The results of the model are shown below in Table 5 with comparative data volumes shown in Figure 6. One day was chosen as the time unit because it allows for a reasonably accurate estimate over multiple orbits, thereby reducing fluctuations in data generation due to non-uniform geography. The data volumes shown are data outputs for the instruments. A 10% overhead amount was added to account for metadata, calibration data, cloud map, and packetization. This average downlink data volume per day of 5,097 Gbits was used to size the communication system.

As the data flow diagram (Figure 5) shows, the model takes into account the satellite orbital parameters, the desired sensor configuration (swath width, trackwise and swathwise pixel resolution), spectral resolution (number of bands), intensity resolution (bits per band), compression ratio, and a model of Earth’s geography. Geography is important since the sensors’ data output rates will vary according to the sensors’ target (Land/Shallow Water (LSW), Cloud, or Ocean). To determine the percentage of time the sensors will view the features, a map was overlaid onto a Satellite Orbit Analysis Program (SOAP) model of the satellite developed by Francois Rogez at JPL. The map determined the satellite’s sensor mode – whether it should record with LSW or Ocean resolutions. By running this model over a 19-day period at 5minute resolution (as to account for the entire repeat cycle), the percentage of time the satellite would collect data from the designated regions (LSW or Ocean) was identified.

Communications Architecture Selection The main driver in the selection of HypsIRI’s communication architecture was the large data downlink rate required. Because the HyspIRI mission aims to provide consistent, lowrisk operations, redundancy and heritage were considered to be major selection factors. Cost and masses of various architectures were also considered. Table 7 shows a comparison between the architectures considered.

Since the locations of clouds are unpredictable, a statistical average was used to predict the time the sensor would be viewing clouds. With an 11 AM local descending node, the expected cloud cover is approximately 30%. Estimating a 2-in-3 detection success rate for pixels that are cloudy, we estimate 20% of LSW areas will be labeled as ‘cloud’ [2]. Note

Ka band frequencies (between 18-40 GHz) are often used in communications satellites. These frequencies are chosen be8

Figure 5. Simplified operation of the data flow model

Table 5. Data sum over one day of normal science operations. The added 10% overhead accounts for the required added metadata.

Table 6. Link Budget for X-band to Svalbard GS.

cause of their ability to provide high gain as well as high data rates with smaller physical structures. The Ka band was ruled to be unfeasible because most ground stations do not support

Ka band, and the use of Ka would mean introducing a new standard to multiple ground stations while modem upgrades to the exisiting X-band receiving systems were ruled more 9

un-upgraded ground stations and increase the data envelope should the upgraded ground stations (at 800 Mbps) fail. Additionally, many high-latitude ground stations such as Svalbard and Fairbanks, Alaska [19], [20], [21] are compatible with the X-band frequencies.

Power Required Data Rate Data Margin Cost

X-band 200 W 800 Mbps 28.5% $9.5M

TDRSS 320 W 300 Mbps 13.3% $11.2M

LaserComm 40-60 W 6 Gbps 23.7% $20-30M

Table 7. Comparison of the different communications architectures considered. Based on these characteristics, an X-band-based architecture was baselined for the mission.

X-band Architecture Details Figure 6. Comparative sizes of the data volumes. Note that LSW (VSWIR) and LSW/Cloud (TIR) represent 98% and 97% of the overall data volume, respectively.

High data rate X-band heritage—Two LEO imaging satellites were benchmarked to find a starting point for the design of the X-band communications system. The first is the DigitalGlobe WorldView-1 satellite, launched in September of 2007. Using two 400-Mbps X-band radios, it achieves an overall data rate of 800 Mbps [18]. The other is the GeoEye-1 Satellite, which will be launched in September of 2008, which is specified to be capable of 740 Mbps and 150 Mbps [22].

cost-effective. “TDRSS is a communication signal relay system which provides tracking and data acquisition services between low earth orbiting spacecraft and NASA/customer control and/or data processing facilities” [16]. TDRSS, while not supporting as high rates as can be achieved by X-band, is capable of continuously operating for long link times (>20 minutes per orbit). TDRSS was investigated due to its long link times and its convenient ground station location at White Sands. However, using TDRSS with the current return link data rate would require more power from HyspIRI as well as more complicated structural requirements due to the need for a larger antenna to meet Effective Isotropic Radiative Power (EIRP) requirements on the HyspIRI-to-TDRSS-side.

The DigitalGlobe WorldView-1 is of special interest because of its data rate. The satellite is specified to collect 331 gigabits of data per orbit, which is extremely close to the 343 gigabits per orbit expected to be collected by HyspIRI mission [18]. Spacecraft Segment for X-band—To achieve 800 Mbps transmission, the satellite would use two 400 Mbps radios simultaneously. To accomplish this at the same band, the signals would be linearly and perpendicularly polarized [1]. The WorldView-1 satellite, which operates at 800 Mbps, uses the L-3 Communications T-722 X-band radio, which operates between 8 and 8.4 GHz. The radio has an output power of 7.5 W with an OQPSK modulation scheme, which increases bitrate from BPSK without introducing error associated with higher-order PSK [23]. To determine if this system would be feasible for HyspIRI, a link budget was created, as can be seen in Table 6. After determining the desired acquisition location, the link geometry (see Table 8) was used to conduct link budget sizing using a sample link budget from David Hansen at JPL. The budget was set up to produce a minimum link margin of 6 dB with the available 7.5 W that is specified by the T-722 radio [24]. The budget is for one radio (each radio requiring data rates of 400 Mbps); since the other will be identical but perpendicularly polarized, the budget applies for both radios.

LaserComm is a relatively new technology under development in labs, including JPL. LaserComm has the potential to transmit data at extremely fast rates (up to 10 Gbps), making it attractive to the HyspIRI mission [17]. While the link speed has been demonstrated, the link rate does not seem to be supportable due to the data rate capabilities of the Solid State Recorder (SSR). Faster storage media must be developed before the satellite can support such a high data rate. It also remains the most expensive of the communications system options. X-band was deemed to be the best solution overall for HyspIRI. The X-band system is estimated to be the least expensive of the three options while having excellent heritage in the DigitalGlobe WorldView-1 satellite both for data rate and data volume [18]. X-band also provides HyspIRI with the ability to reduce the data rate (as GeoEye-1 does) to utilize

As this rough link budget shows, Svalbard ground station is theoretically capable of 10 dB more gain than is necessary to 10

GS Link Elevation Max Slant Distance Off-Nadir Gimballing Angle

10◦ above horizon 1984 km 64◦

Table 8. Link Geometry for 623 km. Note the Gimballing angle must be 64 degrees; this has been ruled feasible with a 0.25m-dia parabolic antenna as noted in Table 6.

make the link. Therefore, there is room to reduce the power, decrease the size of the antenna, or use smaller receiver antennas on the ground (and gain access to more ground stations) while maintaining a robust link. Physically, the radio system will require precise gimballing and rotational control to align the polarization with the receiver antenna at the ground station. The SA-200HP bus that is baselined for the mission has sufficient spacecraft pointing knowledge to supply for the 0.5◦ pointing error assumed in the link budget. The mass of the X-band system is shown below in Table 9. Item Radio Antenna and Gimballing S-band Rx Coax Switch Total Mass

Figure 7. Each radio will downlink a single “File Bundle at a time.

Mass 3.9 kg each 2.5 kg 2.5 kg 0.4 kg 12.2 kg

these stations, no station supports data rates higher than 150 Mbps while most support approximately 100 Mbps [21].

Table 9. Mass Budget for X-band Architecture.

Therefore, to use the 800 Mbps link, the ground stations must be upgraded from their current speed with new electronics and data storage. For the purposes of communications system sizing, it was assumed that with a combination of ground stations, it would be possible to provide an average of 10 minutes of downlink time per orbit. This could be accomplished by upgrading two to three high latitude ground stations.

Because the two radios will be downlinking from the same pool of onboard data, it is necessary to create a file splitting scheme to downlink the data (see Figure 7). Since the instrument compression will be reset every several lines for error containment, the data that is processed will be separated by geographical segments. These segments will each correspond to a “file bundle” which contains the instrument data, calibration data, and cloud map data associated with that segment’s geographical location. Metadata in the files will be used to designate the location the file bundle correlates to. Each file bundle will be assigned to a radio so that at any one moment, two file bundles are transmitted together. During transmission, the next file bundle will then be queued for the next available radio. Since each file bundle contains metadata about its location, the data can be pieced back together at the ground station.

As a contingency option, it should be feasible to decrease the data rate of HyspIRI’s communications system to 100 Mbps to utilize more ground stations for additional coverage time and data volume, or as a backup in case one of the high-rate ground stations is unavailable. Data Volume for X-band Communications— At 800 Mbps transfer rate with an average of 10 minutes of contact time per orbit, the X-band system will be able to downlink 7,125 Gb while the satellite should generate 5,097 Gb on average. This scenario yields an overall data margin of 28.5%.

Ground Segment for X-band— Multiple NASA ground stations support X-band communications with current antennas. These ground stations include high-latitude ground stations Svalbard, Fairbanks, TrollSat, McMurdo, TERSS, and many others. These ground stations have dual-polarizing parabolic antennas, which would be necessary to support the downlink radio frequency structure. While this band is supported at

Cost of X-band Communications—The estimated cost for Xband communications is shown in Table 10. Satellite communications costs are taken from the last TeamX report conducted in 2007. Ground station costs are estimates after consulting Joseph Smith at JPL [25]. Dedicated data line costs were estimated at $1M each over 3 years. 11

Table 10. Estimated cost of the X-band system.

5. G ROUND S TATION S ELECTION As mentioned previously, there are several ground stations that are compatible with X-band communications frequencies. It was necessary to determine how many and which ground stations should be used in order to accommodate the HyspIRI mission’s large data downlink volume. Analysis Methods Twenty ground stations were considered in this analysis. It was assumed that the ground stations would be upgraded to accommodate the 800 Mbs data rate required by the HyspIRI mission. A driving factor in the ground station selection process was the amount of onboard data storage available, currently 1.6Tb based on Team X sizing[1]. A Satellite Orbit Analysis Program (SOAP) scenario was created to provide the downlink durations over the entire 19-day period for each instrument. PERL scripts (created by Francois Rogez at JPL) translated the downlink durations into the amount of data being stored onboard the spacecraft based on the data downlink and uplink rates.

Table 11. Target Values of Objective Function Attributes

Four scenarios were created by varying the values of the weighting coefficients used in Equation 1 (see Table 12). The scenarios were run using one, two and three ground stations. The ground station combination that minimized the objective function was then chosen as the baseline configuration. Results

While the onboard storage capabilities of the spacecraft limited the possible ground station options, other factors also influenced the selection of the ground station configuration. To account for these factors, a decision making tool was created. With respect to the onboard storage, the decision making tool took into account the peak volume, the slope (volume vs. time), and the average volume of the data stored over the 19 day repeat cycle. In addition to these characteristics, the configurations were also evaluated based on the cost associated with upgrading the ground stations, the assumed reliability, and the political control of each ground station. The objective function is shown in Equation 1. The T subscript represents the target and the A subscript represents the actual value. The objective function target values are given below in Table 11. objf unc =

T −M axP eakA α M axP eak M axP eakT T −SlopeA +β SlopeSlope T AvgV alT −AvgV alA +χ AvgV alT +δ(Reliability − 1) +(P oliticalControl − 1) alGS +φ numGS−AvgV AvgV alGS

It was found that the HyspIRI mission’s data volume is too large to be accommodated by just one ground station. For the scenarios generated using just one ground station, the amount of data stored onboard the spacecraft increased linearly over the 19 day collection cycle, clearly violating the 1.6Tb onboard data storage constraint. The amount of data collected by the HyspIRI mission is also too large to be accommodated by two ground stations: again, the amount of data stored onboard increases almost linearly over the 19 day repeat cycle. With three ground stations, however, there are several feasible ground station combinations for this mission. Of the 20 possible ground station combinations, there are 11 solutions where the onboard data storage volume varies sinusoidally (see Figure 8). Note that the trial numbers correspond to the ground station numbers shown in Table 13. However, once the 1.6Tb onboard data storage volume constraint is included in this analysis, there are only two feasible ground station configurations (see Figure 9). The configurations that meet the onboard data storage requirements are Fairbanks, McMurdo and TrollSat and Fairbanks, Svalbard, and TrollSat.

(1)

where: α+β+χ+δ++φ=1

(2) 12

Figure 8. Close-up Time series of amount of data stored onboard for scenarios using three ground stations upgraded to 800 Mbps.

Figure 9. Three ground stations upgraded to 800Mbps solutions with 1.6Tb onboard storage constraint.

13

Table 12. Objective Function Coefficient Scenarios

More advanced HyspIRI mission design conducted at JPL by Team X has now expanded the available onboard data storage to 3Tb.

Sensitivity Analysis Given that only two combinations of three ground stations do not exceed the onboard data storage limits, it is necessary to understand the sensitivity of our results to the onboard data storage limits. This was accomplished by varying the weighting coefficients of the objective function (see Table 12). Table 14 shows the results for each of the scenarios.

6. S CIENCE A NALYSIS In order to more fully understand the requirements associated with the data throughout the mission, it was necessary to define data levels both on-board the spacecraft and on the ground. By defining the form and scientific significance of the data at each stage, a greater understanding of the data system’s role was gained. Additionally, by gaining a greater understanding of HyspIRI’s data products, it was possible to identify missions with which HyspIRI could produce complementary science products.

Table 14 shows that, for every scenario, TrollSat was included in the best ground station combination. Fairbanks was the second-most favored ground station with McMurdo and Svalbard tied for third place. Sweden and Tasmania were never chosen for the best combination. This does not mean that they were not feasible solutions. However, with the given weighting functions, they were never chosen as the best solution.

Data Level Definition The optimal ground station configuration from Scenario 2 (Fairbanks, McMurdo, and Trollsat) was baselined for this mission due to its emphasis on the 1.6Tb maximum onboard data storage constraint.

Definitions for data levels follow (see Figure 10). Level 0 Raw Data: HyspIRI will collect two types of raw data: the science data from the two sensors and the calibration data. As the sensors collect photons, they will output analog signals resulting from the light intensity levels as well as the sensors’ material properties. The calibration data is obtained through lunar, blackbody, solar, and dark level calibration. Lunar and blackbody calibration will be conducted by satellite pointing. Solar calibration will be conducted by reflecting solar light off the instrument cover. Dark level calibration will be conducted by closing the instrument cover. Ephemeris data will be obtained through the onboard GPS and star-sensors. The calibration data will be needed on-board the spacecraft in order to conduct the preliminary calibration required for cloud detection.

Preliminary analysis was conducted into the sensitivity of the system to missing a downlink opportunity. This analysis was conducted for the ground link scenario experienced on the second day of the spacecraft’s 19 day repeat cycle. Also, this analysis does not look at the possibility of loosing contact with several different ground stations over the course of a day. Preliminary results are shown in Table 15 and a more detailed discussion can be found in [26]. It can be seen that virtually all of the cases considered exceed the 1.6Tb of available onboard storage. More work needs to be done in this area to adequately address the risk of missing downlink opportunities throughout the mission.

Level 1A Roughly Calibrated to Reflectance Data:

Table 14. Objective Function Results

Table 13. Ground Station Reference Numbers

14

Ground Station Fairbanks Trollsat Svalbard

Number of Passes Missed 1 All (8) 1 All (10) 1 All(10)

Max Required On-board Storage 2 Tb 2.7 Tb 1.5 Tb 3.7 Tb 1.7 Tb 3.8 Tb

Table 15. Amount of storage required when downlink opportunities are missed. These preliminary results were computed on the second day of the spacecraft’s 19 day repeat cycle. All passes missed indicates that all the passes over the ground station were missed for a given day.

Figure 10. The data level definitions are presented. The chart flows from left to right, top to bottom through levels 0, 1A, 1B, 1C, 2, 3 and 4. These visual examples belong past mission imagery. Visual examples for levels 1A, 2, and 3 belong to AVIRIS data and calibration charts, while visual examples for Level 4 belongs to ASTER data. Once the data is converted to digital number values, it will be calibrated using the dark level calibration values to correct for sensor offsets due to material flaws. During dark level calibration the sensors should not detect any signal, thus any signals from the sensor indicate an erroneous offset. This dark-level-calibrated data must then be converted from sensor readings to spectral radiance values. This conversion process takes inputs from solar and satellite zenith and azimuth angles, radiometric information (stored onboard in coefficient appendices), and top-of-atmosphere effects (scattering). The data will be tagged with universal collection time, and latitude and longitude.

then generate reflectance values. This conversion must be completed onboard because reflectance allows for the detection of clouds for the VSWIR instrument. Level 1B Compressed, Packetized Data: Level 1B data consists of compressed, packetized data. Cloudy pixels undergo a lossy compression and a lossless compression while non-cloudy land-and-shallow-water pixels undergo only lossless compression. Ocean pixels will undergo lossy compression. Because there is little scientific value gained from VSWIR images of clouds, cloud maps are generated to enable higher compression for clouds. In the case of the TIR, all land-and-

Radiance values that have been atmospherically corrected can 15

shallow-water data, regardless of cloud presence, will be subject to lossless compression. Compressed Level 1A data is then packaged alongside the ephemeris metadata to become Level 1B data, which is downlinked to ground stations. Packetization is done such that communications errors causing the loss of one packet will not affect the data integrity of other packets.

sions, and ground-based measurements. These will permit the creation of high-impact science products. It is possible that the HyspIRI mission could collaborate with the Aerosol and Cloud Ecosystems (ACE) mission, which is also motivated by the NRC Decadal Survey. By collaborating with ACE – a mission scheduled launch between 2013 and 2016 – more accurate predictions of aquatic ecosystem responses to acidification and organic budget changes can be obtained from disturbance effects on distribution, biodiversity, and functioning of ecosystems (HyspIRI science products) as well as ocean CO2 sinking and acidity changes (ACE science products). Also, correlation between ecosystem changes and ocean organic measurements can be studied by combining disturbance effects on distribution, biodiversity, and functioning of ecosystems (HyspIRI science products) and oceanic organic material budgets (ACE science product). Through collaboration with other mission’s the impact of HyspIRI’s data products will be increased.

Level 1C Decompressed, Depacketized Data: Once data packets are received on ground, the data is depacketized and decompressed. The packets are then assembled to create full-resolution, continuous data sets as Level 1C data. This data is backed up and will be stored permanently, as is described below. Level 2 Atmospheric Corrected Data: In order to generate Level 2 data, instrument measurement values (previously calibrated) are resampled, remapped, and recalibrated with more sophisticated and accurate methods. Data is also geometrically calibrated and orthorectified. Further corrections may involve cosmic ray compensation and bad pixel replacement. Spectral signatures are identified. A series of high-level information (physical, optical, radiance, and other derived elements) can be viewed at this level. Spectral signatures (derived from reflectance), surface radiance, temperature, emissivity, and composition are available at this level.

HyspIRI’s many science products produced at the various data levels will add to the large volume of data that must be stored once the data is sent to the ground. Additionally, clear definitions of the science products and data levels will enable the creation of an efficient and accessible ground data storage system.

7. DATA S TORAGE AND D ISTRIBUTION Level 3 Mapped Data: Geophysical parameters are mapped onto uniform space grids located in space and time with instrument location, pointing and sampling. Volcanic eruptions, landslides, wildfires, natural resources, drought, soil and vegetation type, will be mapped. Gridded maps are created for each image. Files may be located regionally. A series of overlapping data per orbit may be encountered. All files will be made available, including overlapping files.

Long-term data and science product storage is becoming an increasingly relevent issue in data system design. Afterall, the worth of the mission is based on the accessibility and usability of the data once it has been sent to the ground. The combined data collection rate of HyspIRI’s two instruments exceeds 1.5TB/day and nearly 0.56 PB/year. The raw data ingest rate alone would increase the current Earth Observing System Data and Information System (EOSDIS) daily archive growth by 48%, not including the storage of science products. The HyspIRI Adaptive Data Processing System (HIIAPS) is a proposed data processing architecture based on the Atmospheric Composition Processing System (ACPS) with minor modifications. HIIAPS would handle the challenge of archiving, processing, and distributing such an immense amount of data while keeping costs low.

Level 4 Temporal Dependant Data: As the instruments finish their repeat cycles, they will produce global frames of science products; 5 days per frame for the TIR and 19 days for the VSWIR. With these frames, it is possible to characterize the earth in terms of the science data products every repeat cycle. These frames will be compiled based on collection time, producing a 4-D model of the earth which can be used to view local or global changes as a function of time. This 4-D model would be made available on a monthly, seasonal, or yearly basis.

Lessons Learned from EOSDIS The Earth Observing System (EOS) Data and Information System (EOSDIS) began development and implementation in the early 1990s and has been fully operational since 1999 [27], [28], [29]. EOSDIS is the primary information infrastructure that supports the EOS missions. The responsibilities of EOSDIS could originally be divided into two sectors: mission operations and science operations [30], [27], [28], [31]. Mission Operations handles satellite flight operations, data capture, and initial processing of raw data. Most of the Mission Operations responsibilities are now handled by Earth Science Mission Operations (ESMO) [30], [32]. Sci-

Synergy with Other Missions The HyspIRI mission will produce science products for various monitoring and early warning programs. If combined with science products from other missions, the HyspIRI science products can serve for even greater societal benefit and scientific interest. HyspIRI’s science products can be combined with missions that will be collecting data simultaneously around 2013-2016, as well as with past and future mis16

ence Operations processes, distributes, and archives the data [30], [27], [28], [31]. NASA Integrated Services Network (NISN) mission services handle the data transfer between the mission operations system and the science operations systems [30].

dia gradually as costs and performances improve and as they are needed. To be able to do this in the most cost effective manner a system usage analysis should be conducted in order to properly size the storage and processing systems. For such a system there should be a detailed plan in place for the migration of data from old to new storage media as the older storage media becomes obsolete or nonfunctional [27].

There are several valuable lessons that were learned from the operation and continued evolution of EOSDIS. The Goddard Earth Sciences (GES) Distributed Active Archive Center (DAAC) and Langley Research Center (LaRC) Atmospheric Science Data Center (ASDC) evolutions emphasize reducing operations costs by maintaining only a single processing system, increasing automation through the use of simpler systems, reducing engineering costs through the replacement of COTS (commodity-off-the-shelf) products with commodity products, and increasing accessibility by increasing disk usage by phasing out existing tape based systems [28]. The parallel development of Archive, Next Generation (ANG¯e) and Simple, Scalable, Script-Based, Science Processing Archive (S4PA) at GES DAAC and LaRC ASDC DAAC demonstrate that NASA continues to encourage individual entities, such as the DAACs and Science Investigator-led Processing Systems (SIPSs), to suggest and develop new approaches [28]. A summary of the lessons learned from the development and continued operation of EOSDIS (from “Evolving a Ten Year Old Data System”[28]) are briefly listed below:

Keeping all affected parties involved is especially relevant to the formatting of the data, which will be discussed later. Involvement of these parties combined with continuous smallscale prototyping speeds the generation and integration of new ideas and also serves as a risk management measure to prevent changes being made to the system that would be detrimental to the user community. Involvement of these parties also helps to determine how access to data should be kept online and how that data might be most easily accessed online. HyspIRI Adaptive Processing System HIIAPS (HyspIRI Adaptive Processing System), pronounced “high-apps,” is a proposed stand alone data processing system which would handle the data processing, storage, and distribution requirements of the HyspIRI mission. It would be based on ACPS with only the minor modifications that occur in adapting a previous system to the slightly different needs of another mission. JPL’s Team X states that “Science team has responsibility for all storage and analysis . . . Data is made available (online) to full science team for providing atmospheric corrections and producing products” [1]. An ACPSlike processing system would fulfill these requirements. HIIAPS would be a centralized dedicated SIPS and DAAClike system that would be co-located with the instrument and science teams. The MODIS Adaptive Processing System (MODAPS) has demonstrated that co-locating the science team with the processing center reduces cost by saving time on the development and integration of new algorithms [33]. It is also important to note that HIIAPS would have no commercial or proprietary parts thus evading the issues of software licensing. A diagram of the HIIAPS data processing context is shown in Figure 11.

Having a program requirement for continuous technology assessment establishes a culture of innovation • An on-going, small-scale, prototyping effort within operational elements is important for exploring new technologies and hardware • Involvement by all affected parties is essential to evolution planning • It is necessary to collect and analyze system usage metrics (e.g. request patterns for various data products) in planning for the future • Flexible management processes accommodating innovation, speed, and efficiency are essential for increasing agility in development despite the higher perceived risks • A risk management strategy must be included in planning for evolution • Transitions affecting the user community need to be carefully managed with sufficient advance notification •

All storage and analysis for the HyspIRI mission would be done on site. Level 1C data is archived and distributed at HIIAPS. HIIAPS will also process and archive Level 3 and below data leaving higher level products to academia and the private sector. Data products would be sent directly to the HypsIRI science team and an offsite archive. This offsite archive could be a DAAC or another data processing center that serves as a risk management precaution in the event of a natural disaster at the HIIAPS site. Like MODAPS, HIIAPS would employ data pool technology to save on storage and processing costs. Higher level, smaller data products and recently created products would be kept in the data pool while more cumbersome products like Level 1C data would be produced on demand. Also like previous systems, HIIAPS would provide custom data products in the form of virtual products that are made to order. By building upon the knowledge base

All of these lessons point toward an independently operated system built for continuous evolution. As has been stated before HyspIRI generates extremely large amounts of data. A highly scalable and adaptive system would be the most cost effective approach to handle the processing, archiving, and distribution needs of the mission. The presence of a culture of innovation would facilitate the necessary continuous reassessment of the technology present at HIIAPS. Such technology reassessments would facilitate a smooth continuous integration of new processing and storage components in both the hardware and software realms. Thus initial development costs could be reduced by eliminating the need to purchase all storage media at once and instead purchasing storage me17

plished by distributing the data in a format that is easy to use. A standard format would be helpful because it would guarantee that everybody could read that file format. However, there is no single standard format for geospatial data [38]. NASA’s EOSDIS currently uses HDF-EOS5, which is good for processing but not everyone can easily read the HDF file format [39], [40], [41]. Even though there is no standard some formats are more easily accessible than others. Using a GIS (Geographic Information Systems)-ready format like a flat georeferenced JPEG or PNG image would make HyspIRI data readily accessible to an already large user base and to those who do not use a GIS program. The Open GIS ConsorR tium, Inc. (OGC) is working to solve this formatting issue and towards Open GIS [42], [43]. “Open GIS is the full integration of geospatial data into mainstream information technology. What this means is that GIS users would be able to freely exchange data over a range of GIS software systems and networks without having to worry about format conversion or proprietary data types.”[43] Working with the OGC, or at least staying in touch with their progress, would help to maximize interoperability and accessibility. Based on this HyspIRI should have at least two standard data formats. One would be HDF-EOS5 because that is the standard NASA EOS format and HIIAPS can draw on pervious NASA experience for it. The second would be a GIS ready format like a georeferenced JPEG or PNG to increase accessibility to HyspIRI data.

Figure 11. HIIAPS Data Processing Flow

of previous processing systems, HIIAPS would both reduce the costs and risk of creating a new processing center.

This section looked at how HyspIRI might fit into the EOSDIS framework, several previous processing systems, and touched on accessibility and interoperability concerns. HyspIRI should take advantage of the present EOSDIS infrastructure. However, HIIAPS should be a stand alone processing center so that HyspIRI’s large data volume does not overtax the existing EOSDIS system. There are three heritage systems for HIIAPS: MODAPS, OMIDAPS, and ACPS, each an the evolution of the next. There has been a steady shift from tape-based storage to disk-based storage over the years. Thus HIIAPS should use a disk-based storage system. In addition to storage and processing, distribution is also a concern for the HyspIRI mission. HyspIRI data should be made available at no charge to as many people as possible in a manner accessible to as many people as possible. Some of the challenges are choosing a data format, reaching non-remote sensing scientists, and providing the data in an easy to find manner. Using Google Earth TM and supplying HyspIRI data products in more than one format will help to solve this challenge.

Storage Media: Disk vs. Tape Initially, EOSDIS used tape-based storage. Now it is slowly transitioning to disk based storage [28], [31], [34]. Tape is easily transportable, has faster read/write speeds, but it is sequential access storage and a nearline media [35], [36], [37]. On the other hand, disk has a higher storage capacity per unit, makes the data easily accessible online, experiences more frequent technology upgrades, but has higher electrical costs and is more prone to error [35], [36], [37]. Disk storage is also more convenient to use because it is random access and data can be kept online (as opposed to nearline or offline) all the time, thus shortening data retrieval times. It was been shown that disk storage is more expensive in the short term, but in the long term the cost difference between disk and tape systems seems to converge [29]. This combined with the fact that ACPS, the Ozone Monitoring Instrument Data Processing System (OMIDAPS), and MODAPS are all disk based storage leads to the conclusion that HIIAPS should also be an entirely disk based system, especially with the Kryder’s Law (disk storage densities double every 23 months) taken into account [38].

8. C ONCLUSION An end-to-end data system has been described for the HyspIRI mission, a high data volume Earth-observing mission. With the HyspIRI mission, ambitious science objectives are driving the development of a cutting-edge data handling system, both on-board the spacecraft and on the ground.

Data Formats As the previous section states it is imperative that HyspIRI data is made available to as many non-remote sensing scientists as possible. Part of achieving this goal will be accom18

The HyspIRI mission utilizes innovative techniques to both reduce the amount of data that must be transmitted to the ground and accommodate the required data volume on the ground. The infrastructure and techniques developed by this mission will open the door to future high data volume science missions. The designs presented here are the work of the authors and may differ from the current HyspIRI mission baseline.

[11] C. Burges, “A tutorial on support vector machines for pattern recognition,” Data Mining and Knowledge Discovery, vol. 2, pp. 121–167, 1998. [12] M. Klemish, “Fast lossless compression of multispectral imagery, internal jpl document,” October 2007. [13] F. Rizzo, “Low-complexity lossless compression of hyperspectral imagery via linear prediction,” p. 2, IEEE Signal Processing Letters, IEEE, 2005.

ACKNOWLEDGMENTS

[14] R. Roosta, “Nasa - jpl: Nasa electronic parts and packaging program.” http://nepp.nasa.gov/docuploads/3C8F70A32452-4336-B70CDF1C1B08F805/JPL%20RadTolerant%20FPGAs%20for%20Space%20Applications.pdf, December 2004.

This research was carried out at the Jet Propulsion Laboratory, California Institute of Technology, and was sponsored by the Space Grant program and the National Aeronautics and Space Administration.

[15] I. Xilinx, “Xilinx : Radiation-hardened virtex-4 qpro-v family overview.” http://www.xilinx.com/support/documentation/ data sheets/ds653.pdf, March 2008.

R EFERENCES [1]

K. Warfield, T. V. Houten, C. Heeg, V. Smith, S. Mobasser, B. Cox, Y. He, R. Jolly, C. Baker, S. Barry, K. Klassen, A. Nash, M. Vick, S. Kondos, M. Wallace, J. Wertz Chen, R. Cowley, W. Smythe, S. Klein, L. Cin-Young, D. Morabito, M. Pugh, and R. Miyake, “Hyspiri-tir mission study 2007-07 final report, internal jpl document,” TeamX 923, Jet Propulsion Laboratory, California Institute of Technology, 4800 Oak Grove Drive, Pasadena, CA 91109, July 2007.

[2]

R. O. Green, “Hyspiri summer 2008 overview,” 2008. Information exchanged during presentation.

[3]

S. Hook, 2008. Information exchanged during meeting discussion. July 16th.

[4]

R. O. Green, “Measuring the earth with imaging spectroscopy,” 2008.

[5]

“Moore’s law: Made real by intel innovation.” http://www.intel.com/technology/mooreslaw/index.htm.

[6]

T. Doggett, R. Greeley, S. Chein, R. Castano, and B. Cichy, “Autonomous detection of cryospheric change with hyperion on-board earth observing-1,” Remote Sensing of Environment, vol. 101, pp. 447–462, 2006.

[7]

[8]

[9]

[16] G. S. F. Center, “Tdrss overview.” http://msp.gsfc.nasa.gov/tdrss/oview.html, 7. [17] H. Hemmati, 07 2008. Information exchanged during meeting about LaserComm. [18] “Worldview-1.” http://www.digitalglobe.com/index.php /86/WorldView-1, 2008. [19] “Svalbard ground station, norway.” http://www.aerospacetechnology.com/projects/svalbard/, 7 2008. [20] “Satellite tracking ground http://www.asf.alaska.edu/stgs/, 2008.

station.”

[21] R. Flaherty, “Sn/gn systems overview,” tech. rep., Goddard Space Flight Center, NASA, 7 2002. [22] “Geoeye-1 fact sheet.” http://launch.geoeye.com/launchsite/about/fact sheet.aspx, 2008. [23] L-3 Communications/Cincinnati Electronics, Space Electronics, 7500 Innovation Way, Mason, OH 45040, T-720 Transmitter, 5 2007. PDF Spec Sheet for the T720 Ku-Band TDRSS Transmitter.

R. Castano, D. Mazzoni, N. Tang, and T. Dogget, “Learning classifiers for science event detection in remote sensing imagery,” in Proceedings of the ISAIRAS 2005 Conference, 2005.

[24] L-3 Communications/Cincinnati Electronics, Space Electronics, 7500 Innovation Way, Mason, OH 45040, T-722 X-Band, 7 2007. PDF Spec Sheet for the T-722.

S. Shiffman, “Cloud detection from satellite imagery: A comparison of expert-generated and autmatically-generated decision trees.” ti.arc.nasa.gov/m/pub/917/0917 (Shiffman).pdf.

[25] J. Smith, 07 2008. Information exchanged during meeting about GDS. [26] J. Carpena-Nunez, L. Graham, C. Hartzell, D. Racek, T. Tao, and C. Taylor, “End-to-end data system design for hyspiri mission.” Jet Propulsion Laboratory, Education Office, 2008.

M. Griggin, H. Burke, D. Mandl, and J. Miller, “Cloud cover detection algorithm for eo-1 hyperion imagery,” Geoscience and Remote Sensing Symposium, 2003. IGARSS ’03. Proceedings. 2003 IEEE International, vol. 1, pp. 86–89, July 2003.

[27] J. Behnke, T. Watts, B. Kobler, D. Lowe, S. Fox, and R. Meyer, “Eosdis petabyte archives: Tenth anniversary,” Mass Storage Systems and Technologies, 2005. Proceedings. 22nd IEEE / 13th NASA Goddard Conference on, pp. 81–93, April 2005.

[10] V. Vapnik, Advances in Kernel Methods: Support Vector Learning. MIT Press, 1999. 19

[28] M. Esfandiari, H. Ramapriyan, J. Behnke, and E. Sofinowski, “Evolving a ten year old data system,” Space Mission Challenges for Information Technology, 2006. SMC-IT 2006. Second IEEE International Conference on, pp. 8 pp.–, July 2006.

[42] “Welcome to the ogc http://www.opengeospatial.org/, 2008.

[43] “Open gis: Gis lounge - geographic information systems.” http://gislounge.com/open-gis/.

[29] S. Marley, M. Moore, and B. Clark, “Building costeffective remote data storage capabilities for nasa’s eosdis,” Mass Storage Systems and Technologies, 2003. (MSST 2003). Proceedings. 20th IEEE/11th NASA Goddard Conference on, pp. 28–39, April 2003.

Christine M. Hartzell received her B.S. in Aerospace Engineering for Georgia Institute of Technology with Highest Honors in 2008. She is currently a PhD student at the University of Colorado at Boulder, where she is researching the impact of solar radiation pressure on the dynamics of dust around asteroids. She has spent two summers working at JPL on the data handling system for the HyspIRI mission, with particular emphasis on the cloud detection algorithm development and instrument design.

[30] “Earth science data and information system (esdis) project.” http://esdis.eosdis.nasa.gov/index.html. [31] M. Esfandiari, H. Ramapriyan, J. Behnke, and E. Sofinowski, “Evolution of the earth observing system (eos) data and information system (eosdis),” Geoscience and Remote Sensing Symposium, 2006. IGARSS 2006. IEEE International Conference on, pp. 309–312, 31 2006Aug. 4 2006. [32] “Earth science mission operations http://eos.gsfc.nasa.gov/esmo/.

(esmo).” Jennifer Carpena-Nunez received her B.S. in physics in 2008 from the University of Puerto Rico, where she is currently a PhD student in Chemical Physics. Her research involves field emission studies of nanostructures and she is currently developing a field emission setup for further studies on nanofield emitters. The summer of 2008 she worked at JPL on the HyspIRI mission. There, she was responsible for the science analysis of the data handling system, specifically defining the data level and processing, and determining potential mission collaborations.

[33] E. Masuoka and M. Teague, “Science investigatorled global processing for the modis instrument,” Geoscience and Remote Sensing Symposium, 2001. IGARSS ’01. IEEE 2001 International, vol. 1, pp. 384–386 vol.1, 2001. [34] M. Esfandiari, H. Ramapriyan, J. Behnke, and E. Sofinowski, “Earth observing system (eos) data and information system (eosdis) — evolution update and future,” Geoscience and Remote Sensing Symposium, 2007. IGARSS 2007. IEEE International, pp. 4005–4008, July 2007. [35] D. McAdam, “The evolving role of tape in the data center,” The Clipper Group: Explorer, December 2006. [36] “Sun microsystems announces first one terabyte tape storage http://www.sun.com/aboutsun/pr/200807/sunflash.20080714.2.xml, July 2008.

website.”

world’s drive.”

Lindley C. Graham is currently a junior at the Massachusetts Institute of Technology, where she is working towards a B.S. in Aerospace Engineering. She spent last summer working at JPL on the data handling system for the HyspIRI mission, focusing on developing a data storage and distribution strat-

[37] “Panasas — welcome.” http://www.panasas.com/. [38] R. Domikis, J. Douglas, and L. Bisson, “Impacts of data format variability on environmental visual analysis systems.” http://ams.confex.com/ams/pdfpapers/119728.pdf.

egy.

[39] “Why did nasa choose hdf-eos as the format for data products from the earth observing system (eos) instruments?.” http://hdfeos.net/reference/Info docs/SESDA docs/NASA chooses HDFEOS.php, July 2001. [40] R. E. Ullman, “Status and plans for hdfeos, nasa’s format for eos standard products.” http://www.hdfeos.net/hdfeos status/ HDFEOSStatus.htm, July 2001. [41] “Hdf esdis project.” http://hdf.ncsa.uiuc.edu/projects/esdis/index.html, August 2007. 20

David M. Racek is a senior working toward a B.S. in Computer Engineering at Montana State University. He works in the Montana State Space Science and Engineering Laboratory where he specializes in particle detector instruments and circuits. He spent last summer working at JPL on compression algorithms for the HyspIRI mission.

advanced scientific software for Earth and space science modeling with an emphasis on high performance computing and finite element adaptive methods. Additionally he is leading efforts in development of smart payload instrument concepts. He has given 32 national and international keynote/invited talks; published in numerous journals, conference proceedings, and book chapters. He is a member of the editorial board of the journal Scientific Programming, the IEEE Technical Committee on Scalable Computing, a Senior Member of IEEE, recipient of the JPL Lew Allen Award, and a NASA Exceptional Service Medal.

Tony S. Tao is currently a junior honor student at the Pennsylvania State University, working towards a B.S. in Aerospace Engineering and a Space Systems Engineering Certification. Tony works in the PSU Student Space Programs Laboratory as the project manager of the OSIRIS Cube Satellite and as a systems engineer on the NittanySat nanosatellite, both of which aim to study the ionosphere. During his work at JPL in the summer of 2008, Tony worked on the communication and broadcast system of the HyspIRI satellite as well as a prototype Google Earth module for science product distribution. Christianna E. Taylor received her B.S. from Boston University in 2005 and her M.S. at Georgia Institute of Technology in 2008. She is currently pursing her PhD at the Georgia Institute of Technology and plans to pursue her MBA and Public Policy Certificate in the near future. She worked on the ground station selection for the HyspIRI mission during the summer of 2008 and looks forward to working at JPL in the coming year as a NASA GSRP fellow. Hannah R. Goldberg received her M.S.E.E and B.S.E from the Department of Electrical Engineering and Computer Science at the University of Michigan in 2004 and 2003 respectively. She has been employed at the Jet Propulsion Laboratory, California Institute of Technology since 2004 as a member of the technical staff in the Precision Motion Control and Celestial Sensors group. Her research interests include the development of nano-class spacecraft and microsystems. Charles D. Norton is a Principal Member of Technical Staff at the Jet Propulsion Laboratory, California Institute of Technology. He received his Ph.D in Computer Science from Rensselaer and his B.S.E in Electrical Engineering and Computer Science from Princeton University. Prior to joining JPL he was a National Research Council resident scientist. His work covers 21

Lihat lebih banyak...

Comentários

Copyright © 2017 DADOSPDF Inc.