I. INTRODUCTION
In the last decade, with the introduction of Doppler and dual-polarization technology, weather radars have evolved significantly. The accuracy of weather phenomena interpretation has also improved with high spatiotemporal radar observation [1]. Along with these developments in weather radars, radar based severe weather monitoring and forecasting studies have been extensively conducted. Furthermore, national disaster prevention systems and related services have been established. In particular, in regions where loss of life and property damage frequently occur due to severe weather phenomena such as tornados, localized heavy rainfall, and flash floods, the advanced weather radar system plays an important role for real-time weather monitoring and forecasting in an attempt to minimize these damages.
For radar based weather monitoring system, there is growing interest in the real-time mapping of radar observations on the geographical information system (GIS) and the visualization methods of radar data based on spatial interpolation techniques [2-4]. Currently, the method of displaying radar data in the 2D GIS coordinate system has been extensively used for providing weather information from weather radar. However, it is likely that there is a degree of mapping uncertainty because the elevation of the radar beam and other similar factors are not considered in 2D display methods. Furthermore, mapping to the ground observation from above ground proves difficult. To solve these problems, display systems such as constant altitude plan position indicator (CAPPI) and radar product processing system (PPS), which synthesize radar data observed at various angles of altitude, have been developed[5,6]. However, since scans are done at several altitudes to obtain observations over a volume, mapping accuracy decreases due to time delay and observation time differences during scans. The radar community has also developed techniques to construct 3D volume data by using plan position indicators (PPIs) data at several altitude angles and/or range height indicators (RHIs) data at several azimuths in order to infer the overall size, development, and dissipation of the weather echo[7]. However, such 3D data construction methods have several limitations: real-time work and calculation is complicated. In addition, these methods are not able to show clutter and beam shading region intuitively.
In this paper, we propose a GIS-based high-resolution realistic radar data display method via 3D radar beam radiation modeling. This method can provide users the stereoscopic observation of weather phenomena changes and a way in which to easily discriminate occultation region that occurs from beam blockage. Clutter and shadowing resulted in terrain blockage are especially problematic for radar based quantitative precipitation estimation (QPE) and precipitation pattern recognition[8]. Ground clutters can be misinterpreted as heavy precipitation and therefore overestimations of QPE may occur, whereas QPE underestimations may occur at beam occultation areas after terrain blockage. The proposed display technique that adopts GIS and 3D radar beam radiation modeling can provide users exact locations of clutter and shadowing in order to prevent misinterpretation of precipitation rate or patterns.
The paper is organized as follows: section 2 reviews the process of data generated from the weather radar and section 3 describes the proposed technique in detail. Section 4 demonstrates the test results via simulation of the proposed technique. Finally, section 5 summarizes the proposed method and details the advantages
II. BACKGROUND OF WEATHER RADAR DATA PROCESSING
Lately, the most important technologies in weather radar are dual polarization and networked radar sensing. The X-band radar network has become more popular for urban and mountainous applications with low-level coverage and high spatiotemporal resolution. Single-polarization Doppler radars, which use only one polarization such as horizontal or vertical direction, output the parameters reflectivity (Z), Doppler velocity (V), and spectrum width (W) as part of radar products. Meanwhile, dual-polarization radars, in addition to Z, V and W, obtain dual-polarization products such as differential reflectivity (Zdr), differential propagation phase (PhiDP), copolar correlation coefficient (RhoHV), and specific differential phase (Kdp) [9]. By obtaining such dual-polarization radar variables, precipitation types (rain, snow, hail, etc.), which are difficult to discriminate by using single-polarization radar observations, can be distinguished and the amount of rainfall can be measured more accurately. Figure 1 shows the flow of data processing and data formation of a typical dual-polarization radar. From the process shown in Figure 1(a), the various radar variables mentioned above can be obtained via signal processing and filtering of in-phase and quadrature signals; horizontal polarization signals Ih, Qh and vertical polarization signals Iv, Qv, which are obtained from the signals emitted while the radar antenna rotates. Figure 1(b) shows an example of the reflectivity (Z) among the 1D radar variables, which is dependent on observation distance, formed in real time through the process shown in Figure 1(a). By performing coordinate change and spatial interpolation of radar variables, which are represented in the 1D ray shown in Figure 1(b), one PPI or one RHI radar image can be constructed over each azimuth or elevation.
III. PROPOSED 3D RADAR DATA VISUALIZATION
When the radar beam encounters terrain or buildings, propagating radar signal power decreases according to the degree of occultation. Showing the blockage region is challenging in a real time radar display system. The proposed technique displays 3D radar data in real time on the GIS platform and can reveal clutter and shadowing region resulted in terrain blockage intuitively. The proposed technique maps highly accurate 3D radar data over the signal-processed 1D radar variables obtained from a dual-polarization Doppler radar. The technique displays the 3D radar data in combination with 2D and 3D GIS data in order to analyze the precipitation observation accurately. Figure 2 demonstrates the process of carrying out the proposed technique.
To display 3D radar data in real time, as shown in Figure 2, the 1D variables obtained from the radar, which is the information of each ray, is first extracted while the information of the header of the ray and of the radar variables are processed separately. For 3D displaying, the information of the header of the ray and variables used are defined in Table 1.Figure 3 shows an example of a radar beam shape model using each variable in Table 1. Figure 3(a) demonstrates the radar beam shape at the azimuthal direction (a), whereas Figure 3(b) expresses the beam shape at elevation direction (e) when the radar is observed from the side. Figure 3(c) shows an example of reflectivity profile along a ray, which is obtained in real time from a dual-polarization radar. Here, the radar beam shape is assumed to be circular at the propagating plane, and the beam width (w) and the observation radius (r) have the same values as shown in Figs. 3(a) and (b). Normalized angle values between 0 and 1 are used as the azimuth of a and the elevation angle of e to make it easy to coordinate them with various GIS coordinate systems. Range gate resolution is determined by the radar spatial resolution h as indicated in Table 1 and calculated as h=r/nGates(m).
The proposed technique estimates the shape of the real radar beam, as shown in Figure 3, to display realistic radar data and to model the object of the 3D ray into a horn shape using the parameters in Table 1.
To simplify analyzing data for an observer’s purpose, the object can be formed into horns of various shapes such as plain, circular, rectangular, and hexagonal. Figure 4(a) shows the mimetic diagram of a 3D beam with circular shape using the radar variables in Table 1. Here, the object position that will be seen on GIS is calculated with the radar’s coordinate position, azimuth (a), and the elevation angle of (e), according to the GIS coordinate system using TM (transverse Mercator) or Lat/Lon. In the proposed technique, a 2D array is formed with respect to n≥1 numbers of rays, following which the beam object in Figure 4(a) is classified into upper and lower surfaces based on ground surface. Then, texture mapping is conducted as in Figure 4(b). Finally, radar variables are represented in the 3D beam object. Additionally, the proposed technique is designed to determine the expression method and range of 3D radar data, according to the legend and visual threshold (th) of each radar variable.
IV. SIMULATION RESULT
To simulate the proposed 3D displaying technique of real-time weather radar data, data obtained from the X-band dual-polarization radar of the Korean Institute of Civil Engineering and Building Technology (KICT) and the API of the Open GL[10] 3D graphic library were used. For the GIS coordination, a 2D image map based on a precise TM coordinate system was also used. Figure 5 shows various 3D radar data displays that were constructed using the proposed technique. Figure 5(a) is an example of the 3D object formed using 8 rays and (b)-(d) indicate beam displays with plain, hexagon, and circular shape, respectively.
For real-time observation of massive time-series radar data in real radar observation environments, it is common to construct a dedicated optical communication network between the radar signal processing server and the monitoring system. However, since it is difficult to use the weather radar operating in the field in real time to evaluate the real-time usability of the proposed method, the simulation is performed in an emulation environment where a dedicated network is assumed. Using the previously observed radar moment data, we confirmed that 3D builds and real-time updates are performed without any delays for the transmitted rays obtained by emulating at a rate of 2 Rays/sec. for 720 ray profiles/ 360 azimuth degrees.
The proposed technique enables the weather observer to control the corresponding radar variable threshold based on the characteristics of each radar variable, making it easy to analyze the displayed data. Figure 6 shows the outcome of the threshold change in CZ (corrected reflectivity) in dBZ unit.
Target reflectivity caused by ground clutters in a quality controlled radar image is eliminated via GCF (ground clutter filtering) using Doppler velocity. By using the proposed visualization technique, we can know the GCF filtered areas intuitionally from 3D expressed radar beams on the digital elevation modeled GIS map. Figure 7 shows the terrain masking influence of radar beams so that an observer can draw a fine distinction between GCF filtered and no-echo area.
From figure 7(a) and (b) we can possibly know that the propagations of quality controlled radar beams at low elevations were blocked by mountains. Accordingly, this technique can facilitate some post processing, such as precipitation correction about GCF filtered areas, by observing expressed beams on the 3D GIS Map.
IV. CONCLUSION
This paper proposes a realistic 3D weather radar data display technique with higher spatiotemporal resolution, which is based on the integration of 3D image processing and GIS interaction. This method is focused on stereoscopic visualization, while conventional radar data display methods are based on flat or two-dimensional interpretation. Furthermore, using the proposed technique, the atmospheric change at each moment can simultaneously be observed three-dimensionally at various geological locations. Simulation results indicate that 3D display of weather radar data can be performed in real time. One merit of the proposed technique is that it can provide intuitive understanding of the influence of beam blockage by topography. Through an exact matching each 3D modeled radar beam with 3D GIS map, we can find out the terrain masked areas. Accordingly, the method facilitates the precipitation correction from QPE underestimation caused by GCF. It can also be expected that more accurate short-term forecasting will be possible using stereoscopic observation of weather phenomena changes. Furthermore, we expect that our visualization technique can help people who want to learn the radar system to understand its concept and principle.