Skip to main content
Log in

Object detection for blind inspection of industrial products based on neural architecture search

  • Published:
Journal of Intelligent Manufacturing Aims and scope Submit manuscript

Abstract

Object detection is a key technology to realize the blind inspection of industrial products. To improve the automation degree of building deep convolutional neural networks (CNNs) for object detection and further improve the detection accuracy, this paper proposes an improved neural architecture search method using exclusive-OR (XOR)-based channel feature fusion. First, an XOR-based channel fusion module is designed; it can fuse the feature mapping of different scales at the channel level in the case of multibranch access complementarily. Then, an improved cell pruning strategy is proposed to efficiently prune the connections between cells by setting the architecture parameters of the candidate operations to 0 s, which are in the alignment layers of the subsequent cells. The cell pruning strategy can directly search the multibranch CNN models and narrow the neural network architectures’ gap between the search stage and the evaluation stage. The experimental results show that the proposed method takes approximately 0.75 GPU days to search the optimal neural network on a dataset including six classes for blind inspection of industrial products, and the mean average precision (mAP) is approximately 99.1% on a test dataset, which is higher than those of state-of-the-art methods, e.g., DenseNAS and CSPDarknet53.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8

Similar content being viewed by others

Data availability

The dataset collected in this study cannot be publicly shared at the moment due to the sensitive information involved in the product nameplate. We apologize for any inconvenience caused and will reassess the possibility of releasing the dataset in the future while ensuring the appropriate measures are in place to safeguard sensitive information.

Notes

  1. https://github.com/tzutalin/labelImg/.

References

  • Bochkovskiy, A., Wang, C., & Liao, H. M. (2020). YOLOv4: optimal speed and accuracy of object detection. Computer Vision and Pattern Recognition. https://doi.org/10.48550/arXiv.2004.10934

    Article  Google Scholar 

  • Çelik, A., Küçükmanísa, A., Sümer, A., Çelebi, A. T., & Urhan, O. (2022). A real-time defective pixel detection system for LCDs using deep learning based object detectors. Journal of Intelligent Manufacturing., 33, 985–994. https://doi.org/10.1007/s10845-020-01704-9

    Article  Google Scholar 

  • Chang, J., Zhang, X., Guo, Y., Meng, G., Xiang, S. & Pan, C. (2019) DATA: Differentiable ArchiTecture Approximation. In NeurIPS (pp. 2905–2920).

  • Chen, Q., Wang, Y., Yang, T., Zhang, X., Cheng, J. & Sun, J. (2021) You Only Look One-level Feature. In 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 13034–13043).

  • Chen, L. C., Zhu, Y., Papandreou, G., Schroff, F., & Adam, H. (2018). Encoder-decoder with atrous separable convolution for semantic image segmentation. Computer Vision and Pattern Recognition. https://doi.org/10.48550/arXiv.1802

    Article  Google Scholar 

  • Chu, X., Zhou, T., Zhang, B., & Li, J. (2020). Fair DARTS: Eliminating unfair advantages in differentiable architecture search. Machine Learning. https://doi.org/10.48550/arXiv.1911.12126

    Article  Google Scholar 

  • Du, X. z., Lin, T. Y., Jin, P. c., Ghiasi, G., Tan, M., Cui, Y., Le, Q. V. & Song, X. (2020) SpineNet: Learning Scale-Permuted Backbone for Recognition and Localization. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 11589–11598).

  • Elsken, T., Metzen, J. H., & Hutter, F. (2019). Neural Architecture Search: A Survey. The Journal Of Machine Learning Research., 20(1), 1997–2017. https://doi.org/10.48550/arXiv.1808.05377

    Article  Google Scholar 

  • Fang, J. m., Sun, Y. z., Zhang, Q., Li, Y., Liu, W. & Wang, X. (2020) Densely Connected Search Space for More Flexible Neural Architecture Search. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 10625–10634).

  • He, K., Zhang, X., Ren, S. & Sun, J. (2016) Deep Residual Learning for Image Recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 770–778).

  • Howard, A. G., Sandler, M., Chu, G., Chen, L.-C., Chen, B., Tan, M., Wang, W., Zhu, Y., Pang, R., Vasudevan, V., Le, Q. V. & Adam, H. (2019) Searching for MobileNetV3. In 2019 IEEE/CVF International Conference on Computer Vision (ICCV) (pp. 1314–1324).

  • Kang, M. & Han, B. (2020) Operation-Aware Soft Channel Pruning using Differentiable Masks. In Int Conf on Machine Learning (pp. 5122–5131).

  • Kong, Y., Han, S., Li, X., Lin, Z. & Zhao, Q. (2020) Object detection method for industrial scene based on MobileNet. In the 12th International Conference on Intelligent Human-Machine Systems and Cybernetics (pp. 79–82).

  • Liu, Z., Li, J., Shen, Z., Huang, G., Yan, S. & Zhang, C. (2017) Learning Efficient Convolutional Networks through Network Slimming. In 2017 IEEE International Conference on Computer Vision (pp. 2755–2763).

  • Liu, H. X., Simonyan, K., & Yang, Y. (2018). DARTS: Differentiable Architecture Search. Machine Learning. https://doi.org/10.48550/arXiv.1806.09055

    Article  Google Scholar 

  • Liu, S., Qi, L., Qin, H., Shi, J. & Jia, J. (2018b) Path Aggregation Network for Instance Segmentation. In 2018b IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 8759–8768).

  • Meng, Z., Gu, X., Liang, Y., Dong, X., & Chunguo, W. (2021). Deep Neural Architecture Search: A Survey. Journal of Computer Research and Development (china)., 58(1), 22–33. https://doi.org/10.7544/issn1000-1239.2021.20190851

    Article  Google Scholar 

  • Pham, H., Guan, M. Y., Zoph, B., Le, Q. V. & Dean, J. (2018) Efficient Neural Architecture Search via Parameter Sharing. In Int Conference on Machine Learning 4095–4104

  • Ren, Z., Fang, F., Yan, N., & Wu, Y. (2021). State of the Art in Defect Detection Based on Machine Vision. International Journal of Precision Engineering and Manufacturing-Green Technology., 9, 661–691. https://doi.org/10.1007/S40684-021-00343-6

    Article  Google Scholar 

  • Renq., He, K. m., Girshick, R. B. & Sun, J, S. (2015). Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Transactions on Pattern Analysis and Machine Intelligence., 39, 1137–1149. https://doi.org/10.1109/TPAMI.2016.2577031

    Article  Google Scholar 

  • Simon, N., Friedman, J. H., Hastie, T., & Tibshirani, R. (2013). A Sparse-Group Lasso. Journal of Computational and Graphical Statistics., 22, 231–245. https://doi.org/10.1080/10618600.2012.681250

    Article  Google Scholar 

  • Simonyan, K., & Zisserman, A. (2015). Very deep convolutional networks for large-scale image recognition. Computer Vision and Pattern Recognition. https://doi.org/10.48550/arXiv.1409.1556

    Article  Google Scholar 

  • Sun, P., Zhang, R. f., Jiang, Y., Kong, T., Xu, C., Zhan, W., Tomizuka, M., Li, L., Yuan, Z., Wang, C. & Luo, P. (2021) Sparse R-CNN: End-to-End Object Detection with Learnable Proposals. In 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 14449–14458).

  • Tan, M. x., Chen, B., Pang, R., Vasudevan, V. & Le, Q. V. (2019) MnasNet: Platform-Aware Neural Architecture Search for Mobile. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 2815–2823).

  • Tang, J., Zhou, H., Wang, T., Jin, Z., Wang, Y., & Wang, X. (2022). Cascaded foreign object detection in manufacturing processes using convolutional neural networks and synthetic data generation methodology. Journal of Intelligent Manufacturing., 34, 2925–2941. https://doi.org/10.1007/s10845-022-01976-3

    Article  Google Scholar 

  • Wan, A., Dai, X. l., Zhang, P. z., He, Z., Tian, Y., Xie, S., Wu, B., Yu, M., Xu, T., Chen, K., Vajda, P. & Gonzalez, J. (2020) FBNetV2: Differentiable Neural Architecture Search for Spatial and Channel Dimensions. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 12962–12971).

  • Wang, J., Sun, K., Cheng, T., Jiang, B., Deng, C., Zhao, Y., Liu, D., Mu, Y., Tan, M., Wang, X., Liu, W., & Xiao, B. (2021). Deep high-resolution representation learning for visual recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence., 43, 3349–3364. https://doi.org/10.1109/TPAMI.2020.2983686

    Article  Google Scholar 

  • Wu, B. c., Dai, X. l., Zhang, P. z., Wang, Y., Sun, F., Wu, Y., Tian, Y., Vajda, P., Jia, Y. & Keutzer, K. (2019) FBNet: Hardware-Aware Efficient ConvNet Design via Differentiable Neural Architecture Search. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 10726–10734).

  • Wu, Y., Liu, A., Huang, Z. w., Zhang, S. & Gool, L. V. (2021) Neural Architecture Search as Sparse Supernet. In AAAI Conference on Artificial Intelligence (pp. 10379–10387).

  • Yang, Y., You, S., Li, H., Wang, F., Qian, C. & Lin, Z. (2021) Towards Improving the Consistency, Efficiency, and Flexibility of Differentiable Neural Architecture Search. In 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 6663–6672).

  • Yao, Q., Xu, J., Tu, W. & Zhu, Z. (2020) Efficient Neural Architecture Search via Proximal Iterations. In AAAI Conference on Artificial Intelligence (pp. 6664–6671).

  • Zhangb., Huang, Z. h., Wang, N. y., Xiang, S. m. & Pan, C, X. (2021). You Only Search Once: Single Shot Neural Architecture Search via Direct Sparse Optimization. IEEE Transactions on Pattern Analysis and Machine Intelligence., 43, 2891–2904. https://doi.org/10.1109/TPAMI.2020.3020300

    Article  Google Scholar 

  • Zhaoq., Zheng, P., Xu, S. t. & Wu, X, Z. (2019). Object Detection With Deep Learning: A Review. IEEE Transactions on Neural Networks and Learning Systems., 30, 3212–3232. https://doi.org/10.1109/TNNLS.2018.2876865

    Article  Google Scholar 

  • Zoph, B. & Le, Q. V. (2017) Neural Architecture Search with Reinforcement Learning. arXiv e-prints, arXiv:1611.01578.

Download references

Acknowledgements

This research was supported in part by the National Natural Science Foundation of China (62166012, 62266015), the Guangxi Natural Science Foundation (2022GXNSFAA035644).

Funding

National Natural Science Foundation of China, 62166012, Lin Huang, 62266015, Tie-jun Yang,Natural Science Foundation of Guangxi Province, 2022GXNSFAA035644, Tie-jun Yang.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Tiejun Yang.

Ethics declarations

Conflict of interests

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Huang, L., Deng, W., Li, C. et al. Object detection for blind inspection of industrial products based on neural architecture search. J Intell Manuf 35, 3185–3195 (2024). https://doi.org/10.1007/s10845-023-02199-w

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10845-023-02199-w

Keywords

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy