去雾
去雾
Abstract— Images captured in a hazy environment usually where I(x) is the hazy image captured by the camera, J(x) is
suffer from bad visibility and missing information. Over many the haze-free image, A is the atmospheric light, and t(x) is the
years, learning-based and handcrafted prior-based dehazing algo- transmission map which can be expressed as t (x) = e−βd(x) ,
rithms have been rigorously developed. However, both algorithms
exhibit some weaknesses in terms of haze removal performance. where β is the scattering coefficient and d(x) is the path length
Therefore, in this work, we have proposed the patch-map-based from the sensor to the object. Then, to acquire the haze-free
hybrid learning DehazeNet, which integrates these two strategies scene, (1) can be reformulated as follows:
by using a hybrid learning technique involving the patch map and
a bi-attentive generative adversarial network. In this method, the I (x) − A
J (x) = + A. (2)
reasons limiting the performance of the dark channel prior (DCP) t (x)
have been analyzed. A new feature called the patch map has been
defined for selecting the patch size adaptively. Using this map, the In (2), the transmission map t(x) and the atmospheric light
limitations of the DCP (e.g., color distortion and failure to recover A evidently are crucial for dehazing. However, this problem is
images involving white scenes) can be addressed efficiently. ill-posed since it is difficult to estimate t(x) and A based on
In addition, to further enhance the performance of the method the information of a single image. Several studies based on
for haze removal, a patch-map-based DCP has been embedded
into the network, and this module has been trained with the handcrafted priors and learning techniques have focused on
atmospheric light generator, patch map selection module, and obtaining solutions for these two crucial variables.
refined module simultaneously. A combination of traditional and The dehazing algorithms based on hand-crafted
learning-based methods can efficiently improve the haze removal priors [2]–[6] have been proposed by observing the difference
performance of the network. Experimental results show that between hazy and haze-free images. Tarel and Hautiere [2]
the proposed method can achieve better reconstruction results
compared to other state-of-the-art haze removal algorithms. estimated the atmospheric veil by using the bilateral filter.
He et al. [3] proposed the dark channel prior (DCP) to
Index Terms— Haze removal, end-to-end hybrid learning sys- estimate the transmission map of a hazy image. Zhu et al. [4]
tem, dark channel prior, patch map, bi-attentive generative
adversarial network. developed the color attenuation prior which can predict the
transmission map efficiently. Berman et al. [5] proposed
I. I NTRODUCTION the non-local transmission map obtained from the haze-line
Authorized licensed use limited to: WUHAN UNIVERSITY OF TECHNOLOGY. Downloaded on January 20,2024 at 08:12:47 UTC from IEEE Xplore. Restrictions apply.
6774 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020
Fig. 1. Dehazing results: The hazy input image is shown in (a), while
(b) shows the corresponding ground truth image. The dehazing results
obtained by the traditional DCP, the proposed patch-map-based DCP, and
(e) the proposed patch-map-based hybrid learning DehazeNet (PMHLD) are
shown in (c), (d), and (e), respectively. Fig. 2. Comparison of the performance of the proposed patch-map-based
hybrid learning DehazeNet method with the other state-of-the-art dehazing
algorithms, in terms of the CIEDE2000 and the SSIM, by applying them
on 1000 images from the RESIDE [14] dataset.
to a loss in the color fidelity in some specific scenarios (e.g.,
the DCP may show color distortion in images containing white
and bright scenes). By contrast, the learning-based methods patch-map-based DCP, and the proposed PMHLD method to a
rarely fail, but the visual results may have some limitations hazy image. Clearly, by using the patch-map-based DCP and
since the learned features may not be bound in the haze-related the hybrid learning strategy, the dehazing performance can be
features. enhanced significantly. In Fig. 2, we compare the performance
Therefore, in this work, we proposed an end-to-end archi- of the proposed method with other dehazing algorithms in
tecture called the patch-map-based hybrid learning DehazeNet terms of the structural similarity (SSIM, where a higher value
(PMHLD), which combines two types of strategies. It lever- corresponds to a better result) and the CIEDE2000 metrics
ages the merits of these two strategies and compensate their (related to color distortion, where a lower value corresponds to
weaknesses. In the proposed method, the DCP and the genera- a better result) on the well-known RESIDE dataset [14]. Fig.2
tive adversarial networks (GANs) are adopted as the backbone. shows that the proposed algorithm can achieve outstanding
First, because the patch size affects the performance of the performance compared to other state-of-the-art methods.
DCP significantly, a new feature called the patch map has been To the best of our knowledge, the proposed architecture is
defined for selecting the patch size for each pixel adaptively. the first haze removal algorithm wherein the patch map has
However, the patch map is a complex feature that is difficult to been adopted and the learning-based and prior-based methods
train. Therefore, the bi-attentive patch map selection network have been integrated into a single network. Experimental
(BAPMS-Net) has been proposed to adaptively determine the results show that the proposed method can not only solve
patch size in each pixel based on the bi-attentive discriminator the color distortion problem but can also improve the result
and a new activation function called the patch map ReLU quality of the haze removal significantly. The following are
(PMReLU). By using the patch map, the transmission map the highlights of this work:
can be estimated with high accuracy and the color distortion 1) An end-to-end dehazing system, called PMHLD, has
can be addressed efficiently. been designed, which integrates the patch-map-based DCP to
Second, to further improve the recovered result, the PMHLD the network. With the leverage of the patch map, the entire
has been proposed which embeds the patch-map-based DCP network can be trained jointly with the estimated atmospheric
into the dehazing network by formulating the dark channel light, patch map, and refined network and the images can
layer with a learnable patch map. Using this layer, the dehazing be recovered. Experimental results show that the proposed
network can train the BAPMS-Net, the atmospheric light esti- PMHLD technique can achieve a better performance in recov-
mation network, and the refined network jointly. Thus, in this ering image as compared to the original PMS-Net.
architecture, the recovery process can be carried out with 2) To address the problem of color distortion in the DCP,
the patch-map-based DCP and the learning-based atmospheric the reasons for the failure of DCP in certain scenarios and
light estimation. poor performance owing to the use of a fixed patch size have
Figure 1 shows an example of the dehazed images obtained been investigated. A new feature called patch map has been
from both a quantitative and a qualitative analysis car- proposed. The patch size can be selected adaptively for each
ried out by applying the traditional DCP, the proposed pixel using this new feature.
Authorized licensed use limited to: WUHAN UNIVERSITY OF TECHNOLOGY. Downloaded on January 20,2024 at 08:12:47 UTC from IEEE Xplore. Restrictions apply.
CHEN et al.: PMHLD: PATCH MAP-BASED HYBRID LEARNING DehazeNet FOR SINGLE IMAGE HAZE REMOVAL 6775
3) To predict the patch map efficiently, a novel patch map B. Haze Removal Based on Learning Strategies
selection network based on the bi-attentive GAN and a new
With the development of learning techniques, several dehaz-
activation function, called the patch map ReLU, has been
ing algorithms based on synthetic data have been proposed.
proposed. With these two proposed modules, the prediction
Tang et al. [7] computed the transmission value for every patch
of the patch map can be improved efficiently.
using the random forest. Cai et al. [8] predicted the transmis-
4) To achieve hybrid learning, the DCP has been embed-
sion value for each patch based on an end-to-end system called
ded into the learning process, and the patch-map-based dark
the DehazeNet. Ren et al. [9] proposed the MSCNN to predict
channel layer with a learnable patch map has been proposed.
the transmission map accurately. Li et al. [10] reformulated
By using this layer, both the traditional and the patch-map-
the haze formation model to combine the atmospheric light
based dark channel operations can be integrated in the network
information and the transmission map for haze-free image
learning. To the best of our knowledge, this is the first work
recovery. Ren et al. [20] proposed the CNN-based haze
carried out toward the development of a trainable dark channel
removal algorithm for videos. Zhang and Patel unskip [21]
layer.
introduced the haze formation model and also utilized GANs
The remainder of the paper is organized as follows.
to achieve clear results. In CVPR 2019 [22], an algorithm
Section II presents a review of several conventional haze
based on the patch map concept is proposed to effectively
removal models and learning techniques. Section III provides
address the color distortion problem in haze removal.
the technical details of the proposed algorithm. Section IV
discusses the experimental results. Finally, Section V provides
a conclusion. C. Generative Adversarial Networks
The architecture of the GAN was proposed by
II. R ELATED W ORK Goodfellow et al. [23] in which the generator and the
discriminator are both trained simultaneously to increase the
A. Haze Removal Based on Handcrafted Priors performance of these two networks. This novel architecture
Handcrafted prior methods are usually based on statistical has been widely adopted in a variety of image processing
analysis and observation. He et al. [3] proposed that the dark tasks, such as super resolution [24], rain or haze removal [25],
channel of a haze-free image in the natural scenario is usually image impainting [26], image translation [27], [28], and text-
close to zero. The DCP is expressed as to-image synthesis [29], [30], [31]. Due to the merits of
this architecture, we adopt the GAN as the backbone of our
proposed network because it is able to learn the patch map
J Dark (x) = mi n mi n J k (x) ∼ = 0, (3) and the recover images efficiently.
k∈{r,g,b} y∈(x)
where Jk (x) is the intensity in the color channel k and (x) is III. P ROPOSED M ETHOD
a local patch with fixed size centered at x. Based on (1) and
the DCP, the transmission map t(x) can be estimated by This section describes the proposed PMHLD network. A
flowchart of the network is given in Fig. 3. The proposed
I k (y) J k (y) network is based on two types of dehazing strategies, namely,
mi n mi n = mi n mi n t (x) the patch-map-based DCP and the hybrid-learning DehazeNet.
y∈(x) k∈{r,g,b} A k y∈(x) k∈{r,g,b} A k
+1 − t (x) ∼
= 1 − t (x). (4) Its architecture can be divided into two parts, namely, the
haze-free image generator and the haze-free image discrim-
Therefore, inator. In the generator part, the transmission map and the
atmospheric light are the two crucial variables that need to be
I k (y) estimated. For predicting the transmission map, the DCP was
t (x) = 1 − ω mi n mi n , (5) selected as the backbone, because it achieves an outstanding
y∈(x) k∈{r,g,b} A k
performance compared to the other handcrafted-prior-based
where ω is a constant used for recovering the haze-free image dehazing methods. However, the results recovered using the
with a high accuracy. DCP may usually exhibit severe color distortion problem.
Among other hand-crafted-prior methods, Tarel and Hau- Therefore, the reasons causing this color distortion need to
tiere [2] computed the atmospheric veil in a hazy image be analyzed. To address this problem, a new feature called
based on a bilateral filter and white balance techniques. the patch map has been proposed. To generate the patch map
Fattal et al. [15] [16] recovered hazy images using the albedo accurately, a bi-attentive patch map selection network has
of the scene and the color-line. Zhu et al. [4] established been designed. For the atmospheric light estimation, which
the color attenuation prior and Berman et al. [5] pro- is another important variable for dehazing, an estimation
posed the concept of haze-line to calculate the transmis- network has been developed to predict it precisely. For the
sion map. Meng et al. [17] applied the boundary constraint haze-free discriminator part, because we hope to achieve the
to remove haze. Chen et al. [18] developed the Gradient reconstructed results to be as close as possible to the clear
Residual Minimization method to reduce the visual artifacts. images, the discriminator architecture has been adopted. The
Zhang et al. [19] applied the maximum reflectance prior to abovementioned techniques are described in detail in the
achieve the nighttime image dehazing. following subsections.
Authorized licensed use limited to: WUHAN UNIVERSITY OF TECHNOLOGY. Downloaded on January 20,2024 at 08:12:47 UTC from IEEE Xplore. Restrictions apply.
6776 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020
Nevertheless, in (6), since Jk (y) is usually unknown, the P M (x) = k, wher e k = arg mi n (E k (x)) , i = 1, . . . , n,
k
transmission value cannot be estimated in the haze removal (10)
process. Based on the examples in Fig. 4, one can notice
that the value of some pixels in the denominator in (6) may where PM(x) is the ground-truth patch map and k is the patch
usually be less than 1. This may lead to the value of t(x) map value at the location x. The maximal patch size n was set
Authorized licensed use limited to: WUHAN UNIVERSITY OF TECHNOLOGY. Downloaded on January 20,2024 at 08:12:47 UTC from IEEE Xplore. Restrictions apply.
CHEN et al.: PMHLD: PATCH MAP-BASED HYBRID LEARNING DehazeNet FOR SINGLE IMAGE HAZE REMOVAL 6777
Authorized licensed use limited to: WUHAN UNIVERSITY OF TECHNOLOGY. Downloaded on January 20,2024 at 08:12:47 UTC from IEEE Xplore. Restrictions apply.
6778 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020
where
to predict the transmission map based on the properties of the pmax
p1 = , β0 = α (th 1 − p1) + pmax . (13)
transmission map. In general, the candidates of the activation γ
Authorized licensed use limited to: WUHAN UNIVERSITY OF TECHNOLOGY. Downloaded on January 20,2024 at 08:12:47 UTC from IEEE Xplore. Restrictions apply.
CHEN et al.: PMHLD: PATCH MAP-BASED HYBRID LEARNING DehazeNet FOR SINGLE IMAGE HAZE REMOVAL 6779
Authorized licensed use limited to: WUHAN UNIVERSITY OF TECHNOLOGY. Downloaded on January 20,2024 at 08:12:47 UTC from IEEE Xplore. Restrictions apply.
6780 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020
Authorized licensed use limited to: WUHAN UNIVERSITY OF TECHNOLOGY. Downloaded on January 20,2024 at 08:12:47 UTC from IEEE Xplore. Restrictions apply.
CHEN et al.: PMHLD: PATCH MAP-BASED HYBRID LEARNING DehazeNet FOR SINGLE IMAGE HAZE REMOVAL 6781
Fig. 12. Visual comparison of the dehazing results obtained by applying the proposed PMHLD algorithm and the other state-of-the-art algorithms to a
synthetic dataset.
Authorized licensed use limited to: WUHAN UNIVERSITY OF TECHNOLOGY. Downloaded on January 20,2024 at 08:12:47 UTC from IEEE Xplore. Restrictions apply.
6782 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020
TABLE I TABLE II
C OMPARISON OF PMHLD AND S TATE - OF - THE -A RT M ODELS W ITH THE Q UANTITATIVE E VALUATION R ESULTS , IN T ERMS OF THE F OUR A SSESS -
N UMBER OF PARAMETERS , FLOP S AND RUNTIME . N OTE T HAT, THE MENT M ETRICS , O BTAINED BY A PPLYING THE PMHLD N ETWORK
R ESULT OF DCPDN I S I NVESTIGATED BASED ON 512×512 I MAGES AND THE O THER S TATE - OF - THE -A RT M ETHODS ON THE T EST B
B ECAUSE T HIS M ETHOD R EQUIRED A F IXED -S IZE I NPUT I MAGES . T HE S YMBOL * D ENOTES THE GAN-BASED M ETHOD
we pick 2400 images from the indoor training set (ITS) and the
outdoor training set (OTS) in RESIDE. We use these images to
train the BAPMS-Net and the hybrid DehazeNet in PMHLD. other state-of-the-art networks on Test B images. In this
For the ablation study, 400 images from the ITS and the OTS experiment, 20 existing dehazing algorithms were applied
have been randomly chosen as the test dataset (called Test A). on the test images for comparison. All these results were
For quantitative evaluation, the synthetic objective testing set obtained by using the codes provided by the authors for the
(SOTS), which consists of 1000 indoor and outdoor images, respective networks. Among the different networks, 10 of
has been used (Test B). We have ensured that none of Test them have published in the past 2 years (i.e., 2018-2020).
images were used in the training process. Further, four assessment metrics, namely, MSE, SSIM, the
peak-to-peak signal to noise ratio (PSNR), and the CIEDE2000
B. Training Detail and Model Complexity Analysis color difference, were used to evaluate the performance of
these networks. The MSE, PSNR, and SSIM metrics are
For the training process, each submodule (i.e., the commonly used for calculating the image quality. CIEDE2000
BAPMS-Net, attention modules and atmospheric light estima- can present the color difference between the ground truth and
tion) were pre-trained separately. Following this, all modules the recovered result. A low value of CIEDE2000 indicates a
were trained together in the fine-tuning process. The Adam low color distortion in the recovered result.
optimizer [55] was used with a learning rate of 10−4 . In each As shown in Table II, our proposed method exhibits the best
epoch, the generator was trained with two iterations and two performance compared to other state-of-the-art algorithms in
discriminators were trained with five iterations separately with terms of all four metrics. The experimental results indicate that
a training batch size of 4. In each epoch, 10% of the images the prior-based methods tend to have a severe color distortion
were cut as the validation set. The PMHLD network was problem as compared to the learning-based method. Note that,
implemented on Tensorflow and run on a workstation with the methods in the upper part of the Table are prior-based,
3.7 GHz CPU, 64G RAM, and Nvidia Titan XP GPUs. while the others are learning-based. Moreover, from the results
In terms of its complexity, the proposed PMHLD network of CIEDE2000, it can be observed that our proposed algorithm
consists of 4.94 × 107 parameters and 2 × 108 FLOPs. exhibits outstanding color preservation (at least 27.8% smaller
The average time required for recovering one image of size value of CIEDE2000) compared to other methods. That is, the
480×640 is 0.076 second. A comparison between the PMHLD proposed method can not only remove the haze efficiently but
network and the other state-of-the-art dehazing networks is can also retain the color information. Therefore, the proposed
shown in Table I. These results have been obtained using the idea of the patch map and the combination of the prior-
same workstation as mentioned above and the code released and learning-based methods are effective for image dehazing.
by the authors. The results show that the complexity of the As compared to the traditional DCP, the proposed PMHLD
proposed method is comparable. network exhibits an improved performance (with MSE values
smaller by 74%, SSIM values larger by 8.1%, 37.8% higher
C. Quantitative Analysis on the Synthetic Dataset PSNR, and a 54% less value of CIEDE 2000 color difference).
Table II shows the quantitative comparison of the results In addition, as compared to the GAN-based methods, including
obtained by applying the proposed PMHLD algorithm and the conditional GAN (cGAN) [51], EPDN [49], Cycle (i.e.,
Authorized licensed use limited to: WUHAN UNIVERSITY OF TECHNOLOGY. Downloaded on January 20,2024 at 08:12:47 UTC from IEEE Xplore. Restrictions apply.
CHEN et al.: PMHLD: PATCH MAP-BASED HYBRID LEARNING DehazeNet FOR SINGLE IMAGE HAZE REMOVAL 6783
Fig. 14. Comparison of the results obtained by applying the proposed PMHLD algorithm and other state-of-the-art dehazing algorithms to real-world images.
CycleDehaze) [56], and the DCPDN [21], our method can in the figure. From the figure, it can be observed that the
achieve superior performance in terms of all the metrics as the proposed PMHLD network can achieve a better performance
proposed method integrates both the learning- and prior-based on hazy images of the real world as compared to the other
strategies into the GAN architecture. networks. It can be clearly seen that although the proposed
Figure 12 shows a visual comparison of the dehazed images model has been trained by a synthetic dataset, the robustness
from the SOTS dataset, from which one can see that the of the model still holds and the haze can be efficiently removed
proposed method exhibits excellent performance on dehazing without damaging the image quality. The results recovered
and does not suffer from the color distortion problem. by the prior-based methods (i.e., Meng, CEP, and DCP) may
suffer from a severe color distortion problem (see the 1st ,
the 3rd , and the 4t h rows for the results corresponding to
D. Dehazed Results on Real World Images these algorithms) even if they can remove haze clearly. In the
Figure 14 shows a collection of several hazy images of case of the learning-based methods, the DCPDN sometimes
the real world, which have been used for evaluation in the exhibits the overexposure problem, and the color fidelity may
previous works on dehazing. The proposed PMHLD network be lost (see the 1st , the 6t h , the 9t h , and the 11t h rows). The
and the other state-of-the-art haze removal algorithms have cGAN network shows good performance but suffers from the
been applied to these images and results have been compared color distortion problem (see the 1st , the 2nd , the 3rd and
Authorized licensed use limited to: WUHAN UNIVERSITY OF TECHNOLOGY. Downloaded on January 20,2024 at 08:12:47 UTC from IEEE Xplore. Restrictions apply.
6784 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020
Fig. 15. Comparison of the dehazing results for white and bright scenes. The 1st column shows the input images, the 2nd column shows the results recovered
by the conventional fixed patch size DCP, the 3rd shows the results recovered by the proposed PMHLD method. The 4th and 5th columns show the enlarged
views of the white and bright portions of the dehazing results obtained in the 2nd and the 3rd columns, respectively.
Authorized licensed use limited to: WUHAN UNIVERSITY OF TECHNOLOGY. Downloaded on January 20,2024 at 08:12:47 UTC from IEEE Xplore. Restrictions apply.
CHEN et al.: PMHLD: PATCH MAP-BASED HYBRID LEARNING DehazeNet FOR SINGLE IMAGE HAZE REMOVAL 6785
Fig. 17. Results of the images recovered by applying (b) DCP [3], (c) PDN [59] and (d) the proposed PMHLD on a haze-free image shown in (a).
TABLE III
Q UANTITATIVE E VALUATION OF THE D EHAZING R ESULTS , IN T ERMS OF THE PI [62] AND FADE [57] FOR R EAL -W ORLD I MAGES
TABLE IV
Q UANTITATIVE MSE E VALUATION FOR A BLATION S TUDY ON T EST A. N OTE T HAT, THE M ODULE C I S THE S AME A RCHITECTURE A PPLIED IN O UR
P REVIOUS W ORK (PMS-N ET [22])
Authorized licensed use limited to: WUHAN UNIVERSITY OF TECHNOLOGY. Downloaded on January 20,2024 at 08:12:47 UTC from IEEE Xplore. Restrictions apply.
6786 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020
Fig. 20. Recovered results by selecting patch size (b) 120, (c) 300 where
the corresponding ground truth is in (a).
Fig. 19. The relation between the average time required for image processing
In this section, the analysis on the selection of the maximum
and the selection of the maximum patch size. patch size has been presented. 500 images were randomly
selected from the RESIDE dataset [14] and were resized to
different scales, i.e., 480 × 640, 240 × 320, 120 × 160 and
TABLE V
60 × 80. Their corresponding patch maps were calculated and
Q UANTITATIVE A NALYSIS FOR THE E FFECTIVENESS OF THE P ROPOSED
PATCH M AP AND THE H YBRID L EARNING S TRATEGY ON THE T EST the dehazed images were recovered from the hazy images
B. N OTE T HAT, THE PATCH M AP BASED DCP I S THE O RIGINAL using the process mentioned in Subsection III-B. A plot
V ERSION OF THE PMS-N ET [22] showing the relation between the selection of the maximum
patch size and the corresponding image quality is given in
Fig. 18. From the figure, it is observed that larger input
images require larger maximum patch sizes to achieve higher
recovered image quality. The recovered image quality will
improve with an increase in the maximal patch size. However,
this trend saturates when the maximum patch size exceeds 120.
The variation of the computation time versus the maximum
TABLE VI patch size is presented in Fig. 19. In this experiment, we cal-
Q UANTITATIVE A NALYSIS FOR THE E FFECTIVENESS OF THE P ROPOSED culated the time consumption only in the patch-map-based
ATMOSPHERIC L IGHT E STIMATION M ODULE
dark channel layer because the BAPMS-Net has the same
architecture for the same input image size. It can be seen that
the time required for processing increases significantly with
the maximum patch size. In this work, we chose a maximum
patch size of 120 since it could balance the trade-off between
the efficiency and the effectiveness.
H. Limitation
an experiment is conducted to prove the superior performance In Fig. 20, the limitation of the proposed PMHLD is
of the proposed PMHLD network on haze-free images. In this presented. Notice that the recovered results using the default
experiment, two dark-channel related methods, including the maximum patch size 120 may suffer from the over-exposed
vanilla DCP [3] and the PDN [59] were selected to test problem for the region which has a large area with high inten-
the performance on clean images. In the case of DCP and
sity (see the window and the neighbor region in Fig. 20(b)).
the PMHLD networks, the prior statistics on haze-free images If we apply an even larger upper bound of the patch size,
are investigated (i.e., the dark channel value = 0). In the the over-exposed problem can be avoided (see Fig. 20(b)).
case of PDN [59], the dark channel-like statistical priors were
However, the computation time may be increased (see the
calculated, i.e., 1 − t ≈ 0. comparison in Fig. 18 and 19). There is a tradeoff between
Similar to the experiments carried out in [47], to investigate the well addressing over-exposed region and the computation
the learned statistical regularities, 500 clear images were time.
tested. The histograms and the accumulation of the dark
channel value and 1−t are presented in Figs. 16(a) and 16(b).
V. C ONCLUSION
From the figure, it can be seen that the proposed network
can learn statistical regularity better than the other DCP-based In this work, a novel dehazing algorithm called the PMHLD
methods. Further, a visual comparison of two examples has network, which combines the handcrafted-prior-based and
been shown in Fig. 17. The results recovered by the vanilla the learning-based techniques, has been proposed. First, the
DCP and PDN algorithms tend to be over-dehazed and exhibit weaknesses of the DCP have been analyzed, and a new
some color distortion. In contrast, the proposed PMHLD feature called the patch map has been developed to improve
network avoids the over-dehazing problem. the dehazing process. To generate the patch map accurately,
Authorized licensed use limited to: WUHAN UNIVERSITY OF TECHNOLOGY. Downloaded on January 20,2024 at 08:12:47 UTC from IEEE Xplore. Restrictions apply.
CHEN et al.: PMHLD: PATCH MAP-BASED HYBRID LEARNING DehazeNet FOR SINGLE IMAGE HAZE REMOVAL 6787
a network called the BAPMS-Net has been designed. More- [17] G. Meng, Y. Wang, J. Duan, S. Xiang, and C. Pan, “Efficient image
over, the patch-related features, an attention map, and a new dehazing with boundary constraint and contextual regularization,” in
Proc. IEEE Int. Conf. Comput. Vis., Dec. 2013, pp. 617–624.
activation function called the PMReLU have been adopted for [18] C. Chen, M. N. Do, and J. Wang, “Robust image and video dehazing
the patch map generator. For the discriminator, a bi-attentive with visual artifact suppression via gradient residual minimization,” in
mechanism has been proposed to make the information on the Proc. Eur. Conf. Comput. Vis., 2016, pp. 576–591.
[19] J. Zhang, Y. Cao, S. Fang, Y. Kang, and C. W. Chen, “Fast haze removal
erroneous area more attentive. To further improve the dehazing for nighttime image using maximum reflectance prior,” in Proc. IEEE
performance, an end-to-end haze removal architecture has Conf. Comput. Vis. Pattern Recognit. (CVPR), Jul. 2017, pp. 7418–7426.
been designed to learn the patch map, the atmospheric light, [20] W. Ren et al., “Deep video dehazing with semantic segmentation,” IEEE
Trans. Image Process., vol. 28, no. 4, pp. 1895–1908, Apr. 2019.
and the transmission map simultaneously by proposing the [21] H. Zhang and V. M. Patel, “Densely connected pyramid dehazing
trainable dark channel layer. Experimental results show that network,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit.,
the proposed dehazing system achieves a high recovered image Jun. 2018, pp. 3194–3203.
quality in both the synthetic and the real-world datasets. [22] W.-T. Chen, J.-J. Ding, and S.-Y. Kuo, “PMS-net: Robust haze removal
based on patch map for single images,” in Proc. IEEE/CVF Conf.
Moreover, the proposed model does not lead to color distortion Comput. Vis. Pattern Recognit. (CVPR), Jun. 2019, pp. 11681–11689.
in images. By employing the proposed techniques of the patch [23] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley,
map, end-to-end haze removal architecture, and bi-attentive S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” in
Proc. Adv. Neural Inf. Process. Syst., 2014, pp. 2672–2680.
GAN, high quality dehazing results can be achieved. [24] C. Ledig et al., “Photo-realistic single image super-resolution using
a generative adversarial network,” in Proc. IEEE Conf. Comput. Vis.
Pattern Recognit. (CVPR), Jul. 2017, pp. 4681–4690.
ACKNOWLEDGMENT [25] H. Zhang and V. M. Patel, “Density-aware single image de-raining using
The authors are grateful to the National Center for a multi-stream dense network,” in Proc. IEEE/CVF Conf. Comput. Vis.
Pattern Recognit., Jun. 2018, pp. 695–704.
High-performance Computing for computer time and facilities. [26] R. A. Yeh, C. Chen, T. Y. Lim, A. G. Schwing, M. Hasegawa-Johnson,
and M. N. Do, “Semantic image inpainting with deep generative
models,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR),
R EFERENCES Jul. 2017, pp. 5485–5493.
[1] S. G. Narasimhan and S. K. Nayar, “Chromatic framework for vision [27] T. Xu et al., “AttnGAN: Fine-grained text to image generation with
in bad weather,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. attentional generative adversarial networks,” in Proc. IEEE/CVF Conf.
(CVPR), vol. 1, 2000, pp. 598–605. Comput. Vis. Pattern Recognit., Jun. 2018, pp. 1316–1324.
[2] J.-P. Tarel and N. Hautiere, “Fast visibility restoration from a single [28] L. Wang, V. Sindagi, and V. Patel, “High-quality facial photo-sketch
color or gray level image,” in Proc. IEEE 12th Int. Conf. Comput. Vis., synthesis using multi-adversarial networks,” in Proc. 13th IEEE Int.
Sep. 2009, pp. 2201–2208. Conf. Autom. Face Gesture Recognit. (FG), May 2018, pp. 83–90.
[3] K. He, J. Sun, and X. Tang, “Single image haze removal using dark [29] J. Yu, Z. Lin, J. Yang, X. Shen, X. Lu, and T. S. Huang, “Generative
channel prior,” IEEE Trans. pattern Anal. Mach. Intell., vol. 33, no. 12, image inpainting with contextual attention,” in Proc. IEEE/CVF Conf.
pp. 2341–2353, Dec. 2011. Comput. Vis. Pattern Recognit., Jun. 2018, pp. 5505–5514.
[4] Q. Zhu, J. Mai, and L. Shao, “A fast single image haze removal algorithm [30] S. Reed et al., “Generative adversarial text to image synthesis,” in Proc.
using color attenuation prior,” IEEE Trans. Image Process., vol. 24, 33rd Int. Conf. Mach. Learn., in Proceedings of Machine Learning
no. 11, pp. 3522–3533, Nov. 2015. Research, vol. 48, M. F. Balcan and K. Q. Weinberger, Eds., Jun. 2016,
[5] D. Berman, T. Treibitz, and S. Avidan, “Non-local image dehazing,” in pp. 1060–1069.
Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2016, [31] Z. Zhang, Y. Xie, and L. Yang, “Photographic text-to-image synthesis
pp. 1674–1682. with a hierarchically-nested adversarial network,” in Proc. IEEE/CVF
[6] F. Yuan and H. Huang, “Image haze removal via reference retrieval and Conf. Comput. Vis. Pattern Recognit., Jun. 2018, pp. 6199–6208.
scene prior,” IEEE Trans. Image Process., vol. 27, no. 9, pp. 4395–4409, [32] K. Gregor, I. Danihelka, A. Graves, D. J. Rezende, and D. Wierstra,
Sep. 2018. “DRAW: A recurrent neural network for image generation,” in Proc.
[7] K. Tang, J. Yang, and J. Wang, “Investigating haze-relevant features in a 32nd Int. Conf. Mach. Learn. Res., in Proceedings of Machine Learning
learning framework for image dehazing,” in Proc. IEEE Conf. Comput. Research, vol. 37, Jul. 2015, pp. 1462–1471.
Vis. Pattern Recognit., Jun. 2014, pp. 2995–3000. [33] V. Mnih, N. Heess, A. Graves, and K. kavukcuoglu, “Recurrent models
[8] B. Cai, X. Xu, K. Jia, C. Qing, and D. Tao, “DehazeNet: An end-to-end of visual attention,” in Proc. Adv. Neural Inf. Process. Syst., 2014,
system for single image haze removal,” IEEE Trans. Image Process., pp. 2204–2212.
vol. 25, no. 11, pp. 5187–5198, Nov. 2016. [34] S. You, R. T. Tan, R. Kawakami, Y. Mukaigawa, and K. Ikeuchi,
[9] W. Ren, S. Liu, H. Zhang, J. Pan, X. Cao, and M.-H. Yang, “Single “Adherent raindrop modeling, detectionand removal in video,” IEEE
image dehazing via multi-scale convolutional neural networks,” in Proc. Trans. Pattern Anal. Mach. Intell., vol. 38, no. 9, pp. 1721–1733,
Eur. Conf. Comput. Vis., 2016, pp. 154–169. Sep. 2016.
[10] B. Li, X. Peng, Z. Wang, J. Xu, and D. Feng, “AOD-Net: All-in- [35] R. Qian, R. T. Tan, W. Yang, J. Su, and J. Liu, “Attentive generative
one dehazing network,” in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), adversarial network for raindrop removal from a single image,” in
Oct. 2017, pp. 4770–4778. Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., Jun. 2018,
[11] Q. Liu, X. Gao, L. He, and W. Lu, “Single image dehazing with pp. 2482–2491.
depth-aware non-local total variation regularization,” IEEE Trans. Image [36] S. Zagoruyko and N. Komodakis, “Wide residual networks,” in Proc.
Process., vol. 27, no. 10, pp. 5178–5191, Oct. 2018. Brit. Mach. Vis. Conf., 2016, pp. 1–15.
[12] W.-T. Chen, S.-Y. Yuan, G.-C. Tsai, H.-C. Wang, and S.-Y. Kuo, “Color [37] C. Szegedy, S. Ioffe, V. Vanhoucke, and A. A. Alemi, “Inception-v4,
channel-based smoke removal algorithm using machine learning for inception-resnet and the impact of residual connections on learning,” in
static images,” in Proc. 25th IEEE Int. Conf. Image Process. (ICIP), Proc. 31st AAAI Conf. Artif. Intell., 2017, pp. 4278–4284.
Oct. 2018, pp. 2855–2859. [38] H. Noh, S. Hong, and B. Han, “Learning deconvolution network for
[13] Y. Li et al., “LAP-net: Level-aware progressive network for image semantic segmentation,” in Proc. IEEE Int. Conf. Comput. Vis. (ICCV),
dehazing,” in Proc. IEEE/CVF Int. Conf. Comput. Vis. (ICCV), Dec. 2015, pp. 1520–1528.
Oct. 2019, pp. 3276–3285. [39] C. Peng, X. Zhang, G. Yu, G. Luo, and J. Sun, “Large kernel matters—
[14] B. Li et al., “Benchmarking single-image dehazing and beyond,” IEEE Improve semantic segmentation by global convolutional network,” in
Trans. Image Process., vol. 28, no. 1, pp. 492–505, Jan. 2019. Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jul. 2017,
[15] R. Fattal, “Single image dehazing,” ACM Trans. Graph., vol. 27, no. 3, pp. 4353–4361.
p. 72, 2008. [40] D. Liu et al., “Densely connected large kernel convolutional network
[16] R. Fattal, “Dehazing using color-lines,” ACM Trans. Graph., vol. 34, for semantic membrane segmentation in microscopy images,” in Proc.
no. 1, p. 13, 2014. 25th IEEE Int. Conf. Image Process. (ICIP), Oct. 2018, pp. 2461–2465.
Authorized licensed use limited to: WUHAN UNIVERSITY OF TECHNOLOGY. Downloaded on January 20,2024 at 08:12:47 UTC from IEEE Xplore. Restrictions apply.
6788 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 29, 2020
[41] G. Huang, Z. Liu, L. Van Der Maaten, and K. Q. Weinberger, “Densely Wei-Ting Chen (Student Member, IEEE) received
connected convolutional networks,” in Proc. IEEE Conf. Comput. Vis. the B.S. degree in electrical computer engineer-
Pattern Recognit. (CVPR), Jul. 2017, pp. 4700–4708. ing from National Chiao Tung University, Hsinchu,
[42] A. L. Maas, A. Y. Hannun, and A. Y. Ng, “Rectifier nonlinearities Taiwan, in 2016. He is currently pursuing the
improve neural network acoustic models,” in Proc. ICML, vol. 30, 2013, Ph.D. degree in electronic engineering with National
p. 3. Taiwan University. His research interests relating to
[43] I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, and A. C. Courville, computer vision, digital image processing, machine
“Improved training of Wasserstein GANs,” in Proc. Adv. Neural Inf. learning, and neural networks. His previous research
Process. Syst., 2017, pp. 5767–5777. on image dehazing and desmoking was published by
[44] Y. Liu, J. Pan, J. Ren, and Z. Su, “Learning deep priors for image dehaz- the CVPR and ICIP.
ing,” in Proc. IEEE/CVF Int. Conf. Comput. Vis. (ICCV), Oct. 2019,
pp. 2492–2500.
[45] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks Hao-Yu Fang received the B.S. degree from
for biomedical image segmentation,” in Proc. Int. Conf. Med. Image National Chiao Tung University and the M.S. degree
Comput. Comput.-Assist. Intervent., 2015, pp. 234–241. from National Taiwan University. His research inter-
[46] K. Simonyan and A. Zisserman, “Very deep convolutional networks ests relating to computer vision, machine learning,
for large-scale image recognition,” 2014, arXiv:1409.1556. [Online]. and neural networks.
Available: http://arxiv.org/abs/1409.1556
[47] J. Zhang and D. Tao, “FAMED-net: A fast and accurate multi-scale
end-to-end dehazing network,” IEEE Trans. Image Process., vol. 29,
pp. 72–84, 2020.
[48] J. Zhang, Y. Cao, Y. Wang, C. Wen, and C. W. Chen, “Fully point-wise
convolutional neural network for modeling statistical regularities in
natural images,” in Proc. ACM Multimedia Conf. Multimedia Conf.
(MM), 2018, pp. 984–992. Jian-Jiun Ding (Senior Member, IEEE) was born in
[49] Y. Qu, Y. Chen, J. Huang, and Y. Xie, “Enhanced Pix2pix dehazing Taiwan in 1973. He received the Ph.D. degree from
network,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. National Taiwan University (NTU), Taipei, Taiwan,
(CVPR), Jun. 2019, pp. 8160–8168. in 2001. In 2006, he become an Assistant Professor
[50] K. Mei, A. Jiang, J. Li, and M. Wang, “Progressive feature fusion at the Department of EE and the Graduate Insti-
network for realistic image dehazing,” in Proc. Asian Conf. Comput. tute of Communication Engineering (GICE), NTU.
Vis. (ACCV), 2018, pp. 203–215. In 2012, he was promoted to an Associate Professor.
[51] R. Li, J. Pan, Z. Li, and J. Tang, “Single image dehazing via conditional In 2017, he was promoted to a Professor. His current
generative adversarial network,” in Proc. IEEE/CVF Conf. Comput. Vis. research areas include time-frequency analysis, lin-
Pattern Recognit., Jun. 2018, pp. 8202–8211. ear canonical transforms, wavelet transforms, image
[52] X. Qin, Z. Wang, Y. Bai, X. Xie, and H. Jia, “FFA-net: Feature fusion processing, image compression, integer transforms,
attention network for single image dehazing,” 2019, arXiv:1911.07559. pattern recognition, face recognition, and machine learning.
[Online]. Available: http://arxiv.org/abs/1911.07559
[53] A. Golts, D. Freedman, and M. Elad, “Unsupervised single image Sy-Yen Kuo (Fellow, IEEE) received the B.S. degree
dehazing using dark channel prior loss,” IEEE Trans. Image Process., in electrical engineering from National Taiwan Uni-
vol. 29, pp. 2692–2701, 2020. versity in 1979, the M.S. degree in electrical and
[54] Z. Xu, X. Yang, X. Li, and X. Sun, “Strong baseline for single computer engineering from the University of Cali-
image dehazing with deep features and instance normalization,” in Proc. fornia at Santa Barbara in 1982, and the Ph.D. degree
BMVC, 2018, p. 5. in computer science from the University of Illinois
[55] D. P. Kingma and J. Ba, “Adam: A method for stochastic opti- at Urbana-Champaign (UIUC) in 1987. He was
mization,” 2014, arXiv:1412.6980. [Online]. Available: http://arxiv. the Dean of the College of Electrical Engineering
org/abs/1412.6980 and Computer Science, NTU, from 2012 to 2015,
[56] D. Engin, A. Genc, and H. K. Ekenel, “Cycle-dehaze: Enhanced Cycle- and the Chairman of the Department of Electrical
GAN for single image dehazing,” in Proc. IEEE/CVF Conf. Comput. Engineering, NTU, from 2001 to 2004. He was a
Vis. Pattern Recognit. Workshops (CVPRW), Jun. 2018, pp. 825–833. faculty member with the Department of Electrical and Computer Engineering,
[57] L. Kwon Choi, J. You, and A. C. Bovik, “Referenceless prediction of The University of Arizona, from 1988 to 1991, and an Engineer with Fairchild
perceptual fog density and perceptual image defogging,” IEEE Trans. Semiconductor and Silvar-Lisco, both in California, from 1982 to 1984.
Image Process., vol. 24, no. 11, pp. 3888–3901, Nov. 2015. He is currently a Distinguished Professor with the Department of Electri-
[58] T. M. Bui and W. Kim, “Single image dehazing using color ellipsoid cal Engineering, National Taiwan University (NTU), Taiwan. He has pub-
prior,” IEEE Trans. Image Process., vol. 27, no. 2, pp. 999–1009, lished 450 articles in journals and conferences, and also holds 22 U.S. patents,
Feb. 2018. 23 Taiwan patents, and 15 patents from other countries. His current research
[59] D. Yang and J. Sun, “Proximal dehaze-net: A prior learning-based deep interests include dependable and secure systems, Internet of Things, and image
network for single image dehazing,” in Proc. Eur. Conf. Comput. Vis. processing.
(ECCV), 2018, pp. 702–717. Dr. Kuo was a member of IEEE Computer Society Board of Governors
[60] W. Ren et al., “Gated fusion network for single image dehaz- from 2017 to 2019. He received the Distinguished Academic Achievement
ing,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2018, Alumni Award from the UIUC Department of Computer Science in 2019,
pp. 3253–3261. the Distinguished Research Award, and the Distinguished Research Fellow
[61] S. Santra, R. Mondal, and B. Chanda, “Learning a patch quality Award from the Ministry of Science and Technology in Taiwan. He was
comparator for single image dehazing,” IEEE Trans. Image Process., also a recipient of the Best Paper Awards in the 1996 International Sym-
vol. 27, no. 9, pp. 4598–4607, Sep. 2018. posium on Software Reliability Engineering and the 1986 IEEE/ACM Design
[62] Y. Blau, R. Mechrez, R. Timofte, T. Michaeli, and L. Zelnik-Manor, Automation Conference, and the US National Science Foundation’s Research
“The 2018 PIRM challenge on perceptual image super-resolution,” in Initiation Award in 1989. He is the Vice President of IEEE Computer Society
Proc. Eur. Conf. Comput. Vis. (ECCV), 2018, pp. 334–355. in 2020.
Authorized licensed use limited to: WUHAN UNIVERSITY OF TECHNOLOGY. Downloaded on January 20,2024 at 08:12:47 UTC from IEEE Xplore. Restrictions apply.