Skip to main content

Advertisement

Log in

Dsa-PAML: a parallel automated machine learning system via dual-stacked autoencoder

  • Original Article
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

Finding a high-performance machine learning pipeline (ML pipeline) for a supervised learning task takes much time. It requires many choices, including preprocessing datasets, selecting algorithms, tuning hyperparameters, and ensembling candidate models. With increasing pipelines arises a combination explosion problem. This work presents a new automated machine learning (AutoML) system called Dsa-PAML to address this challenge by recommending, training, and ensembling suitable models for supervised learning tasks. Dsa-PAML is a parallel automated system based on a dual-stacked autoencoder (Dsa). Firstly, meta-features of datasets and ML pipelines are used to alleviate cold-start recommendation problems. Secondly, a novel dual-stacked autoencoder is used to simultaneously learn the latent features of datasets and ML pipelines, efficiently learning collaborations of both datasets and ML pipelines and recommending suitable ML pipelines for a new dataset. Thirdly, Dsa-PAML can train the recommended ML pipelines on the new dataset in a parallel method, which substantially reduces the time complexity of the proposed method. Finally, a parallel selective ensemble system is embedded into Dsa-PAML. It selects base models from candidate ML pipelines according to their runtime, classification performance, and diversity on the validation set, enhancing Dsa-PAML’s stability for most datasets. Amounts of experiments on 30 UCI datasets show that our approach outperforms current state-of-the-art methods.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

References

  1. Baer JL (1973) A survey of some theoretical aspects of multiprocessing. ACM Comput Surv (CSUR) 5(1):31–80

    Article  Google Scholar 

  2. Baudart G, Hirzel M, Kate K, Ram P, Shinnar A, Tsay J (2021) Pipeline combinators for gradual automl. Advances in Neural Information Processing Systems, 34

  3. Chen X, Wujek B (2020) Autodal: distributed active learning with automatic hyperparameter selection. Proceedings of the AAAI conference on artificial intelligence, 34(04):3537–3544 https://doi.org/10.1609/aaai.v34i04.5759. https://ojs.aaai.org/index.php/AAAI/article/view/5759

  4. Davies A, Veličković P, Buesing L, Blackwell S, Zheng D, Tomašev N, Tanburn R, Battaglia P, Blundell C, Juhász A et al (2021) Advancing mathematics by guiding human intuition with ai. Nature 600(7887):70–74

    Article  Google Scholar 

  5. Dong B, Zhu Y, Li L, Wu X (2020) Hybrid collaborative recommendation via dual-autoencoder. IEEE Access 8:46030–46040

    Article  Google Scholar 

  6. Feldmann J, Youngblood N, Karpov M, Gehring H, Li X, Stappers M, Le Gallo M, Fu X, Lukashchuk A, Raja AS et al (2021) Parallel convolutional processing using an integrated photonic tensor core. Nature 589(7840):52–58

    Article  Google Scholar 

  7. Feurer M, Klein A, Eggensperger K, Springenberg JT, Blum M, Hutter F (2015) Efficient and robust automated machine learning. In: proceedings of the 28th International conference on neural information processing systems, Volume 2, NIPS’15. MIT Press, Cambridge, MA, USA, p 2755-2763

  8. Fusi N, Sheth R, Elibol M (2018) Probabilistic matrix factorization for automated machine learning. Adv Neural Inf Process Syst 31:3348–3357

    Google Scholar 

  9. Guo X, Lin W, Li Y, Liu Z, Yang L, Zhao S, Zhu Z (2020) Dken: deep knowledge-enhanced network for recommender systems. Inf Sci 540:263–277

    Article  MathSciNet  Google Scholar 

  10. Hall M, Frank E, Holmes G, Pfahringer B, Reutemann P, Witten IH (2009) The weka data mining software: an update. ACM SIGKDD Explor Newslett 11(1):10–18

    Article  Google Scholar 

  11. Kaissis GA, Makowski MR, Rückert D, Braren RF (2020) Secure, privacy-preserving and federated machine learning in medical imaging. Nat Mach Intell 2(6):305–311

    Article  Google Scholar 

  12. Komer B, Bergstra J, Eliasmith C (2014) Hyperopt-sklearn: automatic hyperparameter configuration for scikit-learn. In: ICML workshop on AutoML, Citeseer. vol 9, p 50

  13. Kong H, Yan J, Wang H, Fan L (2020) Energy management strategy for electric vehicles based on deep q-learning using bayesian optimization. Neural Comput Appl 32(18):14431–14445

    Article  Google Scholar 

  14. Krogh A, Vedelsby J et al (1995) Neural network ensembles, cross validation, and active learning. Adv Neural Inf Process Syst 7:231–238

    Google Scholar 

  15. Kumar MR, Venkatesh J, Rahman AMZ (2021) Data mining and machine learning in retail business: developing efficiencies for better customer retention. J Ambient Intell Human Comput, pp 1–13

  16. Kuncheva LI, Whitaker CJ (2003) Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy. Mach Learn 51(2):181–207

    Article  Google Scholar 

  17. LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436–444

    Article  Google Scholar 

  18. Li D, Deng L, Cai Z (2021) Design of traffic object recognition system based on machine learning. Neural Comput Appl 33(14):8143–8156

    Article  Google Scholar 

  19. Li L, Jamieson K, De Salvo G, Talwalkar RA, Hyperband A (2016) A novel bandit-based approach to hyperparameter optimization. Computer Vision and Pattern Recognition

  20. Li Y, Jiawei J, Gao J, Shao Y, Zhang C, Cui B (2020) Efficient automatic cash via rising bandits. Proc AAAI Conf Artif Intell 34:4763–4771. https://doi.org/10.1609/aaai.v34i04.5910

    Article  Google Scholar 

  21. Lian G, Wang Y, Qin H, Chen G (2021) Towards unified on-road object detection and depth estimation from a single image. Int J Mach Learn Cybernet, pp 1–11

  22. Lu H, Ma X, Ma M (2021) A hybrid multi-objective optimizer-based model for daily electricity demand prediction considering covid-19. Energy 219:119568

    Article  Google Scholar 

  23. Maher M, Sakr S (2019) Smartml: A meta learning-based framework for automated selection and hyperparameter tuning for machine learning algorithms. In: EDBT: 22nd International conference on extending database technology

  24. Marmolin H (1986) Subjective mse measures. IEEE Trans Syst Man Cybernet 16(3):486–489

    Article  Google Scholar 

  25. Marowka A (2018) Python accelerators for high-performance computing. J Supercomput 74(4):1449–1460

    Article  Google Scholar 

  26. Mohr F, Wever M, Hüllermeier E (2018) Ml-plan: automated machine learning via hierarchical planning. Mach Learn 107(8):1495–1515

    Article  MathSciNet  Google Scholar 

  27. Najafi M, Sadoghi M, Jacobsen HA (2020) Scalable multiway stream joins in hardware. IEEE Trans Knowl Data Eng 32(12):2438–2452. https://doi.org/10.1109/TKDE.2019.2916860

    Article  Google Scholar 

  28. Narciso DA, Martins F (2020) Application of machine learning tools for energy efficiency in industry: a review. Energy Rep 6:1181–1199

    Article  Google Scholar 

  29. Nguyen V, Gupta S, Rana S, Li C, Venkatesh S (2019) Filtering bayesian optimization approach in weakly specified search space. Knowl Inf Syst 60(1):385–413

    Article  Google Scholar 

  30. Olson RS, Moore JH (2016) Tpot: a tree-based pipeline optimization tool for automating machine learning. In: Workshop on automatic machine learning, PMLR. pp 66–74

  31. Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, Blondel M, Prettenhofer P, Weiss R, Dubourg V et al (2011) Scikit-learn: machine learning in python. J Mach Learn Res 12:2825–2830

    MathSciNet  MATH  Google Scholar 

  32. Rakotoarison H, Schoenauer M, Sebag M (2019) Automated machine learning with monte-carlo tree search. In: Proceedings of the twenty-eighth international joint conference on artificial intelligence, IJCAI-19. International joint conferences on artificial intelligence organization. pp 3296–3303, https://doi.org/10.24963/ijcai.2019/457

  33. Rama K, Kumar P, Bhasker B (2021) Deep autoencoders for feature learning with embeddings for recommendations: a novel recommender system solution. Neural Comput Appl 33:1–11

    Article  Google Scholar 

  34. Raschka S (2018) Mlxtend: providing machine learning and data science utilities and extensions to python’s scientific computing stack. J Open Sour Softw 3(24):638

    Article  Google Scholar 

  35. Ribeiro MHDM, dos Santos Coelho L (2020) Ensemble approach based on bagging, boosting and stacking for short-term prediction in agribusiness time series. Appl Soft Comput 86:105837

    Article  Google Scholar 

  36. Rosenfield GH, Fitzpatrick-Lins K (1986) A coefficient of agreement as a measure of thematic classification accuracy. Photogramm Eng Remote Sens 52(2):223–227

    Google Scholar 

  37. de Sá AG, Pinto WJG, Oliveira LOV, Pappa GL (2017) Recipe: a grammar-based framework for automatically evolving classification pipelines. In: European conference on genetic programming, Springer. pp 246–261

  38. Sauter NK, Hattne J, Grosse-Kunstleve RW, Echols N (2013) New python-based methods for data processing. Acta Crystallogr Sect D Biol Crystallogr 69(7):1274–1282

    Article  Google Scholar 

  39. Shao T, Zhang H, Cheng K, Zhang K, Bie L (2021) The hierarchical task network planning method based on monte carlo tree search. Knowl-Based Syst 225:107067

    Article  Google Scholar 

  40. Shi J, Yu T, Goebel K, Wu D (2021) Remaining useful life prediction of bearings using ensemble learning: the impact of diversity in base learners and features. J Comput Inf Sci Eng 21(2):0210074

    Article  Google Scholar 

  41. Sun T, Zhou ZH (2018) Structural diversity for decision tree ensemble learning. Frontiers Comput Sci 12(3):560–570

    Article  Google Scholar 

  42. Tahmasebi H, Ravanmehr R, Mohamadrezaei R (2021) Social movie recommender system based on deep autoencoder network using twitter data. Neural Comput Appl 33(5):1607–1623

    Article  Google Scholar 

  43. Thornton C, Hutter F, Hoos HH, Leyton-Brown K (2013) Auto-weka: combined selection and hyperparameter optimization of classification algorithms. In: Proceedings of the 19th ACM SIGKDD international conference on Knowledge discovery and data mining, pp 847–855

  44. Vanschoren J, Van Rijn JN, Bischl B, Torgo L (2014) Openml: networked science in machine learning. ACM SIGKDD Explor Newslett 15(2):49–60

    Article  Google Scholar 

  45. Wang L, Xie S, Li T, Fonseca R, Tian Y (2021) Sample-efficient neural architecture search by learning actions for monte carlo tree search. IEEE Transactions on Pattern Analysis and Machine Intelligence

  46. Wang Q, Xu W, Zheng H (2018) Combining the wisdom of crowds and technical analysis for financial market prediction using deep random subspace ensembles. Neurocomputing 299:51–61

    Article  Google Scholar 

  47. Wang Y, Wang D, Geng N, Wang Y, Yin Y, Jin Y (2019) Stacking-based ensemble learning of decision trees for interpretable prostate cancer detection. Appl Soft Comput 77:188–204

    Article  Google Scholar 

  48. Wei GW (2019) Protein structure prediction beyond alphafold. Nat Mach Intell 1(8):336–337

    Article  Google Scholar 

  49. Wolpert DH (1992) Stacked generalization. Neural Netw 5(2):241–259

    Article  Google Scholar 

  50. Xiao J, He C, Jiang X, Liu D (2010) A dynamic classifier ensemble selection approach for noise data. Inf Sci 180(18):3402–3421

    Article  Google Scholar 

  51. Yang C, Akimoto Y, Kim DW, Udell M (2019) Oboe: collaborative filtering for automl model selection. In: Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining, pp 1173–1183

  52. Yang C, Fan J, Wu Z, Udell M (2020) Automl pipeline selection: efficiently navigating the combinatorial space. In: proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining, pp 1446–1456

  53. Zhang F, Zhai J, Wu B, He B, Chen W, Du X (2021) Automatic irregularity-aware fine-grained workload partitioning on integrated architectures. IEEE Trans Knowl Data Eng 33(3):867–881. https://doi.org/10.1109/TKDE.2019.2940184

    Article  Google Scholar 

  54. Zhang G, Liu Y, Jin X (2020) A survey of autoencoder-based recommender systems. Frontiers Comput Sci 14(2):430–450

    Article  Google Scholar 

  55. Zhang S, Yao L, Xu X, Wang S, Zhu L (2017) Hybrid collaborative recommendation via semi-autoencoder. In: international conference on neural information processing. Springer, pp 185–193

  56. Zhu B, Han J, Zhao J, Wang H (2020) Combined hierarchical learning framework for personalized automatic lane-changing. IEEE Transactions on Intelligent Transportation Systems

  57. Zhuang F, Zhang Z, Qian M, Shi C, Xie X, He Q (2017) Representation learning via dual-autoencoder for recommendation. Neural Netw 90:83–89

    Article  Google Scholar 

Download references

Acknowledgements

This work was supported by the National Key R&D Program of China under Grant No. 2019B090916002.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xiaofeng Zhou.

Ethics declarations

Conflict of interest

The authors declare that they have no conflicts of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendices

Appendix A: The details of ML pipelines in the MKB

The hyperparameters configurations of the data-preprocessors and the ML classifiers are shown in Tables 11 and 12. The first column is the model’s name and the number of times the associated item appears in the MKB. The second column is hyperparameters and ranges corresponding to the models, where \({\{i, j,...\}}\) is the discrete value; [ij] denotes the closed interval between i and j; (ij) denotes the open interval between i and j.

Table 11 The data-preprocessors and hyperparameters configurations
Table 12 The ML classifiers and hyperparameters configurations

Appendix B: The attributes of the test datasets

See Table 13 and Fig. 11.

Fig. 11
figure 11

The size of test datasets. Here, we can see the size of small datasets is smaller than 100 thousand, the size of medium datasets is smaller than 10 million, and the size of large datasets is larger than 10 million

Table 13 The attributes of the test datasets

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Liu, P., Pan, F., Zhou, X. et al. Dsa-PAML: a parallel automated machine learning system via dual-stacked autoencoder. Neural Comput & Applic 34, 12985–13006 (2022). https://doi.org/10.1007/s00521-022-07119-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-022-07119-2

Keywords

Navigation

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy