AnanthiMVijayakumarK (2021) Stock market analysis using candlestick regression and market trend prediction (CKRM). Journal of Ambient Intelligence and Humanized Computing12(5): 4819–4826.
2.
CholletF (2015) Keras. Available at: https://keras.io (accessed 5 August 2020)
3.
ClevertDAUnterthinerTHochreiterS (2015) Fast and accurate deep network learning by exponential linear units (elus). arXiv preprint arXiv: 1511.07289.
4.
DeySKoshinakaTMotlicekP, et al. (2018) DNN based speaker embedding using content information for text-dependent speaker verification. In: IEEE international conference on acoustics, speech and signal processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018, pp.5344–5348. New York, NY: IEEE.
5.
DuaDGraffC (2019) UCI Machine Learning Repository. Irvine, CA: University of California, School of Information and Computer Science. Available at: http://archive.ics.uci.edu/ml (accessed 5 August 2020)
6.
DuchiJHazanESingerY (2011) Adaptive subgradient methods for online learning and stochastic optimization. Journal of Machine Learning Research12(7): 2121–2159.
7.
GlorotXBordesABengioY (2011) Deep sparse rectifier neural networks. In: Fourteenth international conference on artificial intelligence and statistics, vol. 15. Fort Lauderdale, FL, USA Publisher-PMLR, pp.315–323.
8.
HahnloserRHSarpeshkarRMahowaldMADouglasRJSeungHS (2000) Digital selection and analogue amplification coexist in a cortex-inspired silicon circuit. Nature405(6789): 947–951.
9.
HeKZhangXRenS, et al. (2015) Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In: IEEE international conference on computer vision, Santiago, ChilePublisher: IEEE pp.1026–1034.
10.
HendrycksDGimpelK (2016) Gaussian error linear units (gelus). arXiv preprint arXiv: 1606.08415.
11.
HochreiterS (1991) Untersuchungen zu dynamischen neuronalen Netzen. Diploma, Technische Universität München, 91(1).
12.
HochreiterSBengioYFrasconiP, et al. (2001) Gradient flow in recurrent nets: The difficulty of learning long-term dependencies. In: KolenJFKremerSC (eds) A Field Guide to Dynamical Recurrent Neural Networks. New York, NY: IEEE Press, PP. 237–244.
13.
KadamVJadhavSVijayakumarK (2019) Breast cancer diagnosis using feature ensemble learning based on stacked sparse autoencoders and softmax regression. Journal of Medical Systems43: 263.
14.
KingmaDPBaJ (2014) Adam: A method for stochastic optimization. arXiv preprint arXiv: 1412.6980.
15.
KlambauerGUnterthinerTMayrA, et al. (2017) Self-normalizing neural networks. In: GuyonILuxburgUVBengioSWallachHR. FergusRVishwanathanSGarnettR (eds) Advances in Neural Information Processing Systems. Curran Associates Inc., NY, USA, pp.971–980.
16.
KrizhevskyASutskeverIHintonGE (2012) Imagenet classification with deep convolutional neural networks. In: PereiraFBurgesCJCBottouLWeinbergerKQ (eds) Advances in Neural Information Processing Systems. Curran Associates Inc., NY, USA, pp.1097–1105.
17.
KurianBJyothiV (2021) Breast cancer prediction using an optimal machine learning technique for next generation sequences. Concurrent Engineering29(1): 49–57.
18.
LeCunYBengioYHintonG (2015) Deep learning. Nature521(7553): 436–444.
19.
MaasALHannunAYNgAY (2013) Rectifier nonlinearities improve neural network acoustic models. In: Atlanta, GA, USA, Publisher-PMLR (eds) Proceedings of the 30th international conference on machine learning, ICML, Atlanta, GA.
20.
MangasarianOLStreetWNWolbergWH (1995) Breast cancer diagnosis and prognosis via linear programming. Operations Research43(4): 570–577.
21.
MenonAMehrotraKMohanCK, et al. (1996) Characterization of a class of sigmoid functions with applications to neural networks. Neural Networks9(5): 819–835.
22.
NairVHintonGE (2010) Rectified linear units improve restricted boltzmann machines. In: Proceedings of the 27th international conference on machine learning, ICML, Haifa, Israel.
23.
NwankpaCIjomahWGachaganA, et al. (2018) Activation functions: Comparison of trends in practice and research for deep learning. arXiv preprint arXiv: 1811.03378.
24.
PedamontiD (2018) Comparison of non-linear activation functions for deep neural networks on MNIST classification task. arXiv preprint arXiv: 1804.02763.
25.
QinYWangXZouJ (2018) The optimized deep belief networks with improved logistic sigmoid units and their application in fault diagnosis for planetary gearboxes of wind turbines. IEEE Transactions on Industrial Electronics66(5): 3814–3824.
RamachandranPZophBLeQV (2017b) Searching for activation functions. arXiv preprint arXiv: 1710.05941.
28.
ReddiSJKaleSKumarS (2019) On the convergence of Adam and beyond. arXiv preprint arXiv: 1904.09237.
29.
RoseJDJaspinKVijayakumarK (2021) Lung cancer diagnosis based on image fusion and prediction using CT and PET image. In: PriyaERajinikanthV (eds) Signal and Image Processing Techniques for the Development of Intelligent Healthcare Systems. Springer, Singapore, pp.67–86.
30.
SchaulTAntonoglouISilverD (2013) Unit tests for stochastic optimization. arXiv preprint arXiv: 1312.6055.
31.
SrivastavaNHintonGKrizhevskyA, et al. (2014) Dropout: A simple way to prevent neural networks from overfitting. The Journal of Machine Learning Research15(1): 1929–1958.
32.
StreetWNWolbergWHMangasarianOL (1993) Nuclear feature extraction for breast tumor diagnosis. In: Biomedical image processing and biomedical visualization, 29July1993, vol. 1905, pp.861–870. International Society for Optics and Photonics.
33.
SutskeverIMartensJDahlG, et al. (2013) On the importance of initialization and momentum in deep learning. In: 30th international conference on machine learning (ICML-13), pp.1139–1147.
34.
SzegedyCLiuWJiaY, et al. (2015) Going deeper with convolutions. In: IEEE conference on computer vision and pattern recognition, pp. 1–9, Boston, USA: IEEE
35.
TielemanTHintonG (2012) Divide the gradient by a running average of its recent magnitude. Lecture. COURSERA: Neural networks for machine learning.
WolbergWHStreetWNMangasarianOL (1994) Machine learning techniques to diagnose breast cancer from image-processed nuclear features of fine needle aspirates. Cancer Letters77(2–3): 163–171.
38.
ZeilerMD (2012) ADADELTA: An adaptive learning rate method. arXiv: 1212.5701.
39.
ZhangCWoodlandPC (2015) Parameterised sigmoid and ReLU hidden activation functions for DNN acoustic modelling. In: Sixteenth annual conference of the international speech communication association, Dresden, Germany, 6–10 September 2015.
40.
ZhangCWoodlandPC (2016) DNN speaker adaptation using parameterised sigmoid and ReLU hidden activation functions. In: IEEE international conference on acoustics, speech and signal processing (ICASSP), Shanghai, China, 20–25 March 2016, pp.5300–5304. New York, NY: IEEE.