Application of convolutional neural network to traditional data

•Propose a feature grid-based CNN model, FGCN, on traditional data.•Propose methods of converting instance with form of 1-d vector to feature grid.•The performance of FGCN model has reached the state-of-the-art technique XGBoost.•The positions of features in the grid have little influence on predict...

Full description

Saved in:
Bibliographic Details
Published inExpert systems with applications Vol. 168; p. 114185
Main Authors Zhang, Xiaohang, Wu, Fengmin, Li, Zhengren
Format Journal Article
LanguageEnglish
Published New York Elsevier Ltd 15.04.2021
Elsevier BV
Subjects
Online AccessGet full text

Cover

Loading…
Abstract •Propose a feature grid-based CNN model, FGCN, on traditional data.•Propose methods of converting instance with form of 1-d vector to feature grid.•The performance of FGCN model has reached the state-of-the-art technique XGBoost.•The positions of features in the grid have little influence on prediction accuracy.•Fully connected layers in CNN give little marginal classification performance. Convolutional neural networks (ConvNets) have been applied to various types of data, including image, text, and speech, but not to traditional data. In this study, traditional data are defined as data whose features have no spatial or temporal dependencies but might have statistical correlations. We construct a feature grid-based ConvNet (FGCN) model for classification tasks on traditional data. The FGCN model is composed of two functional parts: The first is used to convert traditional data in the form of a 1-D feature vector into a 1-D, 2-D, or higher-dimensional feature grid; and the second is a ConvNet classifier for the converted data. The experimental results show that the FGCN model performs well; therefore, it is worth considering this model for classification tasks on traditional data.
AbstractList Convolutional neural networks (ConvNets) have been applied to various types of data, including image, text, and speech, but not to traditional data. In this study, traditional data are defined as data whose features have no spatial or temporal dependencies but might have statistical correlations. We construct a feature grid-based ConvNet (FGCN) model for classification tasks on traditional data. The FGCN model is composed of two functional parts: The first is used to convert traditional data in the form of a 1-D feature vector into a 1-D, 2-D, or higher-dimensional feature grid; and the second is a ConvNet classifier for the converted data. The experimental results show that the FGCN model performs well; therefore, it is worth considering this model for classification tasks on traditional data.
•Propose a feature grid-based CNN model, FGCN, on traditional data.•Propose methods of converting instance with form of 1-d vector to feature grid.•The performance of FGCN model has reached the state-of-the-art technique XGBoost.•The positions of features in the grid have little influence on prediction accuracy.•Fully connected layers in CNN give little marginal classification performance. Convolutional neural networks (ConvNets) have been applied to various types of data, including image, text, and speech, but not to traditional data. In this study, traditional data are defined as data whose features have no spatial or temporal dependencies but might have statistical correlations. We construct a feature grid-based ConvNet (FGCN) model for classification tasks on traditional data. The FGCN model is composed of two functional parts: The first is used to convert traditional data in the form of a 1-D feature vector into a 1-D, 2-D, or higher-dimensional feature grid; and the second is a ConvNet classifier for the converted data. The experimental results show that the FGCN model performs well; therefore, it is worth considering this model for classification tasks on traditional data.
ArticleNumber 114185
Author Zhang, Xiaohang
Li, Zhengren
Wu, Fengmin
Author_xml – sequence: 1
  givenname: Xiaohang
  orcidid: 0000-0002-5315-8712
  surname: Zhang
  fullname: Zhang, Xiaohang
  email: zhangxiaohang@bupt.edu.cn
  organization: School of Economics and Management, Beijing University of Posts and Telecommunications, Beijing 100876, China
– sequence: 2
  givenname: Fengmin
  surname: Wu
  fullname: Wu, Fengmin
  email: wufm@bupt.edu.cn
  organization: School of Economics and Management, Beijing University of Posts and Telecommunications, Beijing 100876, China
– sequence: 3
  givenname: Zhengren
  orcidid: 0000-0003-2908-2233
  surname: Li
  fullname: Li, Zhengren
  email: lizhengren@bupt.edu.cn
  organization: School of Economics and Management, Beijing University of Posts and Telecommunications, Beijing 100876, China
BookMark eNp9kEtLxDAUhYMoODP6B1wVXHfMzaNpQJBh8AUDbnQdMmkCrbWpSTqD_97WztrV4d57zuXwLdF55zuL0A3gNWAo7pq1jUe9JpiMC2BQ8jO0gFLQvBCSnqMFllzkDAS7RMsYG4xBYCwW6GHT921tdKp9l3mXGd8dfDtMo26zzg7hT9LRh88s-SwFXdWna6WTvkIXTrfRXp90hT6eHt-3L_nu7fl1u9nlhkqWciNLrIE7g4kFvucMXOmYc5XkhRSSO0Is49QRWok9lbqAYk_BEgOFIdwBXaHb-W8f_PdgY1KNH8LYIirCJKHAS5Cji8wuE3yMwTrVh_pLhx8FWE2gVKMmUGoCpWZQY-h-Dtmx_6G2QUVT287Yqg7WJFX5-r_4LxGzcq4
CitedBy_id crossref_primary_10_4018_IJBCE_301215
crossref_primary_10_3389_frans_2023_1125049
crossref_primary_10_3390_app12052737
crossref_primary_10_1007_s11042_023_17419_x
crossref_primary_10_3390_app11157147
crossref_primary_10_1007_s10489_021_03033_7
crossref_primary_10_1080_15481603_2024_2353982
Cites_doi 10.1038/nature14539
10.1109/TIT.1967.1053964
10.1007/s11263-015-0816-y
10.1198/tech.2003.s770
10.1016/j.eswa.2018.02.029
10.1016/j.eswa.2020.113408
10.1201/9781315139470-8
10.1145/2939672.2939785
10.1016/S0893-6080(98)00116-6
10.1016/j.neucom.2005.12.126
10.1007/BF00116251
10.1109/CVPR.2014.220
10.1109/CVPR.2016.90
10.3115/v1/D14-1181
10.1145/1390156.1390177
10.1007/BF00994018
10.1109/ICCV.2011.6126456
10.1016/j.neucom.2018.09.038
10.1109/CVPR.2018.00813
10.1023/A:1010933404324
10.1088/1749-4699/8/1/014008
ContentType Journal Article
Copyright 2020 Elsevier Ltd
Copyright Elsevier BV Apr 15, 2021
Copyright_xml – notice: 2020 Elsevier Ltd
– notice: Copyright Elsevier BV Apr 15, 2021
DBID AAYXX
CITATION
7SC
8FD
JQ2
L7M
L~C
L~D
DOI 10.1016/j.eswa.2020.114185
DatabaseName CrossRef
Computer and Information Systems Abstracts
Technology Research Database
ProQuest Computer Science Collection
Advanced Technologies Database with Aerospace
Computer and Information Systems Abstracts – Academic
Computer and Information Systems Abstracts Professional
DatabaseTitle CrossRef
Computer and Information Systems Abstracts
Technology Research Database
Computer and Information Systems Abstracts – Academic
Advanced Technologies Database with Aerospace
ProQuest Computer Science Collection
Computer and Information Systems Abstracts Professional
DatabaseTitleList Computer and Information Systems Abstracts

DeliveryMethod fulltext_linktorsrc
Discipline Computer Science
EISSN 1873-6793
ExternalDocumentID 10_1016_j_eswa_2020_114185
S0957417420309192
GroupedDBID --K
--M
.DC
.~1
0R~
13V
1B1
1RT
1~.
1~5
4.4
457
4G.
5GY
5VS
7-5
71M
8P~
9JN
9JO
AAAKF
AABNK
AACTN
AAEDT
AAEDW
AAIAV
AAIKJ
AAKOC
AALRI
AAOAW
AAQFI
AARIN
AAXUO
AAYFN
ABBOA
ABFNM
ABMAC
ABMVD
ABUCO
ABYKQ
ACDAQ
ACGFS
ACHRH
ACNTT
ACRLP
ACZNC
ADBBV
ADEZE
ADTZH
AEBSH
AECPX
AEKER
AENEX
AFKWA
AFTJW
AGHFR
AGJBL
AGUBO
AGUMN
AGYEJ
AHHHB
AHJVU
AHZHX
AIALX
AIEXJ
AIKHN
AITUG
AJOXV
ALEQD
ALMA_UNASSIGNED_HOLDINGS
AMFUW
AMRAJ
AOUOD
APLSM
AXJTR
BJAXD
BKOJK
BLXMC
BNSAS
CS3
DU5
EBS
EFJIC
EFLBG
EO8
EO9
EP2
EP3
F5P
FDB
FIRID
FNPLU
FYGXN
G-Q
GBLVA
GBOLZ
HAMUX
IHE
J1W
JJJVA
KOM
LG9
LY1
LY7
M41
MO0
N9A
O-L
O9-
OAUVE
OZT
P-8
P-9
P2P
PC.
PQQKQ
Q38
ROL
RPZ
SDF
SDG
SDP
SDS
SES
SPC
SPCBC
SSB
SSD
SSL
SST
SSV
SSZ
T5K
TN5
~G-
29G
AAAKG
AAQXK
AAXKI
AAYXX
ABKBG
ABXDB
ACNNM
ADJOM
ADMUD
AFJKZ
AKRWK
ASPBG
AVWKF
AZFZN
CITATION
EJD
FEDTE
FGOYB
G-2
HLZ
HVGLF
HZ~
R2-
RIG
SBC
SET
SEW
WUQ
XPP
ZMT
7SC
8FD
JQ2
L7M
L~C
L~D
ID FETCH-LOGICAL-c394t-c980a15fc02e15b541f8f4ffd9569795f22e453f23d7b39a616b31e2c16c25f13
IEDL.DBID AIKHN
ISSN 0957-4174
IngestDate Fri Sep 13 08:04:59 EDT 2024
Thu Sep 26 18:36:25 EDT 2024
Fri Feb 23 02:48:44 EST 2024
IsPeerReviewed true
IsScholarly true
Keywords Traditional data
Convolutional neural network
Data conversion
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-c394t-c980a15fc02e15b541f8f4ffd9569795f22e453f23d7b39a616b31e2c16c25f13
ORCID 0000-0002-5315-8712
0000-0003-2908-2233
PQID 2492315819
PQPubID 2045477
ParticipantIDs proquest_journals_2492315819
crossref_primary_10_1016_j_eswa_2020_114185
elsevier_sciencedirect_doi_10_1016_j_eswa_2020_114185
PublicationCentury 2000
PublicationDate 2021-04-15
PublicationDateYYYYMMDD 2021-04-15
PublicationDate_xml – month: 04
  year: 2021
  text: 2021-04-15
  day: 15
PublicationDecade 2020
PublicationPlace New York
PublicationPlace_xml – name: New York
PublicationTitle Expert systems with applications
PublicationYear 2021
Publisher Elsevier Ltd
Elsevier BV
Publisher_xml – name: Elsevier Ltd
– name: Elsevier BV
References Bergstra, Bardenet, Bengio, Kégl (b0015) 2011
Wang, X., Girshick, R., Gupta, A., & He, K. (2018). Non-local Neural Networks. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. https://doi.org/10.1109/CVPR.2018.00813.
He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. https://doi.org/10.1109/CVPR.2016.90.
Collobert, R., & Weston, J. (2008). A unified architecture for natural language processing. Proceedings of the 25th International Conference on Machine Learning - ICML ’08, 160–167. https://doi.org/10.1145/1390156.1390177.
Russakovsky, Deng, Su, Krause, Satheesh, Ma, Huang, Karpathy, Khosla, Bernstein, Berg, Fei-Fei (b0130) 2015; 115
Bergstra, J., Komer, B., Eliasmith, C., Yamins, D., & Cox, D. D. (2015). Hyperopt: A Python library for model selection and hyperparameter optimization. Computational Science and Discovery, 8(1). https://doi.org/10.1088/1749-4699/8/1/014008.
Breiman, Friedman, Olshen, Stone (b0030) 2017
Kingma, D. P., & Ba, J. (2014). Adam: A Method for Stochastic Optimization. International Conference on Learning Representations. http://arxiv.org/abs/1412.6980.
LeCun, Cortes, Burges (b0115) 1998
Simonyan, K., & Zisserman, A. (2015). Very deep convolutional networks for large-scale image recognition. 3rd International Conference on Learning Representations, ICLR 2015 - Conference Track Proceedings. https://arxiv.org/abs/1409.1556.
Cortes, Vapnik (b0045) 1995; 20
Schapire (b0135) 1999; 2
Cover, Hart (b0050) 1967; 13
Zhang, Zhang, Chen, Sun, Ma, Yu (b0175) 2019; 323
Quinlan (b0125) 1986; 1
Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., … Rabinovich, A. (2014). GoogLeNet Going Deeper with Convolutions. ArXiv Preprint ArXiv:1409.4842. https://doi.org/10.1109/ICCV.2011.6126456.
Taigman, Y., Yang, M., Ranzato, M., & Wolf, L. (2014). DeepFace: Closing the gap to human-level performance in face verification. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 1701–1708. https://doi.org/10.1109/CVPR.2014.220.
Chen, T., & Guestrin, C. (2016). XGBoost: A scalable tree boosting system. Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 13-17-Augu, 785–794. https://doi.org/10.1145/2939672.2939785.
Dutta, Batabyal, Basu, Acton (b0060) 2020; 159
Shrikumar, A., Greenside, P., & Kundaje, A. (2017). Learning important features through propagating activation differences. 34th International Conference on Machine Learning, ICML 2017, 7, 4844–4866. https://arxiv.org/abs/1704.02685.
Kim, Y. (2014). Convolutional neural networks for sentence classification. EMNLP 2014 - 2014 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference, 1746–1751. https://doi.org/10.3115/v1/d14-1181.
Ziegel (b0180) 2003; 45
Qian (b0120) 1999; 12
Bergstra, Bengio (b0010) 2012; 13
J. Ross Quinlan. (2014). C4.5: Programs for Machine Learning. Elsevier.
Kvamme, Sellereite, Aas, Sjursen (b0105) 2018; 102
Dua, D., & Graff, C. (2017). {UCI} Machine Learning Repository. http://archive.ics.uci.edu/ml.
LeCun, Bengio, Hinton (b0110) 2015; 521
.
Kaggle. (2016). Loan Default Prediction.
Breiman (b0025) 2001; 45
Hox, Bechger (b0070) 1998; 11
Van Der Maaten, Hinton (b0165) 2008
Krizhevsky, Sutskever, Hinton (b0100) 2017; 60
Berger, Della Pietra, Della Pietra (b0005) 1996; 22
Huang, Zhu, Siew (b0075) 2006; 70
10.1016/j.eswa.2020.114185_b0055
Bergstra (10.1016/j.eswa.2020.114185_b0010) 2012; 13
10.1016/j.eswa.2020.114185_b0155
10.1016/j.eswa.2020.114185_b0035
Krizhevsky (10.1016/j.eswa.2020.114185_b0100) 2017; 60
Cortes (10.1016/j.eswa.2020.114185_b0045) 1995; 20
10.1016/j.eswa.2020.114185_b0090
10.1016/j.eswa.2020.114185_b0170
10.1016/j.eswa.2020.114185_b0095
10.1016/j.eswa.2020.114185_b0150
Dutta (10.1016/j.eswa.2020.114185_b0060) 2020; 159
Van Der Maaten (10.1016/j.eswa.2020.114185_b0165) 2008
Bergstra (10.1016/j.eswa.2020.114185_b0015) 2011
Breiman (10.1016/j.eswa.2020.114185_b0030) 2017
Schapire (10.1016/j.eswa.2020.114185_b0135) 1999; 2
10.1016/j.eswa.2020.114185_b0065
Huang (10.1016/j.eswa.2020.114185_b0075) 2006; 70
Qian (10.1016/j.eswa.2020.114185_b0120) 1999; 12
10.1016/j.eswa.2020.114185_b0145
Quinlan (10.1016/j.eswa.2020.114185_b0125) 1986; 1
10.1016/j.eswa.2020.114185_b0080
Hox (10.1016/j.eswa.2020.114185_b0070) 1998; 11
Zhang (10.1016/j.eswa.2020.114185_b0175) 2019; 323
10.1016/j.eswa.2020.114185_b0040
Cover (10.1016/j.eswa.2020.114185_b0050) 1967; 13
Ziegel (10.1016/j.eswa.2020.114185_b0180) 2003; 45
Berger (10.1016/j.eswa.2020.114185_b0005) 1996; 22
10.1016/j.eswa.2020.114185_b0085
10.1016/j.eswa.2020.114185_b0140
10.1016/j.eswa.2020.114185_b0020
Breiman (10.1016/j.eswa.2020.114185_b0025) 2001; 45
LeCun (10.1016/j.eswa.2020.114185_b0115) 1998
Russakovsky (10.1016/j.eswa.2020.114185_b0130) 2015; 115
LeCun (10.1016/j.eswa.2020.114185_b0110) 2015; 521
Kvamme (10.1016/j.eswa.2020.114185_b0105) 2018; 102
References_xml – year: 1998
  ident: b0115
  article-title: The MNIST database of handwritten digits
  publication-title: The Courant Institute of Mathematical Sciences.
  contributor:
    fullname: Burges
– volume: 1
  start-page: 81
  year: 1986
  end-page: 106
  ident: b0125
  article-title: Induction of decision trees
  publication-title: Mach Learn
  contributor:
    fullname: Quinlan
– volume: 2
  start-page: 1401
  year: 1999
  end-page: 1406
  ident: b0135
  article-title: A Brief Introduction to Boosting
  publication-title: IJCAI International Joint Conference on Artificial Intelligence
  contributor:
    fullname: Schapire
– volume: 13
  start-page: 21
  year: 1967
  end-page: 27
  ident: b0050
  article-title: Nearest neighbor pattern classification
  publication-title: IEEE Transactions on Information Theory
  contributor:
    fullname: Hart
– volume: 11
  start-page: 354
  year: 1998
  end-page: 373
  ident: b0070
  article-title: An introduction to structural equation modeling
  publication-title: Family Science Review
  contributor:
    fullname: Bechger
– volume: 45
  start-page: 5
  year: 2001
  end-page: 32
  ident: b0025
  article-title: Random forests
  publication-title: Machine Learning
  contributor:
    fullname: Breiman
– volume: 12
  start-page: 145
  year: 1999
  end-page: 151
  ident: b0120
  article-title: On the momentum term in gradient descent learning algorithms
  publication-title: Neural Networks
  contributor:
    fullname: Qian
– volume: 521
  start-page: 436
  year: 2015
  end-page: 444
  ident: b0110
  article-title: Deep learning
  publication-title: Nature
  contributor:
    fullname: Hinton
– volume: 45
  start-page: 267
  year: 2003
  end-page: 268
  ident: b0180
  article-title: The Elements of Statistical Learning
  publication-title: Technometrics
  contributor:
    fullname: Ziegel
– year: 2008
  ident: b0165
  article-title: Visualizing data using t-SNE
  publication-title: Journal of Machine Learning Research
  contributor:
    fullname: Hinton
– volume: 22
  start-page: 39
  year: 1996
  end-page: 68
  ident: b0005
  article-title: A Maximum Entropy Approach to Natural Language Processing
  publication-title: Computational Linguistics
  contributor:
    fullname: Della Pietra
– volume: 159
  start-page: 113408
  year: 2020
  ident: b0060
  article-title: An efficient convolutional neural network for coronary heart disease prediction
  publication-title: Expert Systems with Applications
  contributor:
    fullname: Acton
– volume: 115
  start-page: 211
  year: 2015
  end-page: 252
  ident: b0130
  article-title: ImageNet large scale visual recognition challenge
  publication-title: International Journal of Computer Vision
  contributor:
    fullname: Fei-Fei
– year: 2017
  ident: b0030
  article-title: Classification and regression trees
  publication-title: Classification and Regression Trees
  contributor:
    fullname: Stone
– volume: 13
  start-page: 281
  year: 2012
  end-page: 305
  ident: b0010
  article-title: Random search for hyper-parameter optimization
  publication-title: The Journal of Machine Learning Research
  contributor:
    fullname: Bengio
– volume: 20
  start-page: 273
  year: 1995
  end-page: 297
  ident: b0045
  article-title: Support-vector networks
  publication-title: Mach Learn
  contributor:
    fullname: Vapnik
– volume: 60
  start-page: 84
  year: 2017
  end-page: 90
  ident: b0100
  article-title: ImageNet classification with deep convolutional neural networks
  publication-title: Advances in Neural Information Processing Systems
  contributor:
    fullname: Hinton
– volume: 323
  start-page: 37
  year: 2019
  end-page: 51
  ident: b0175
  article-title: Recent advances in convolutional neural network acceleration
  publication-title: Neurocomputing
  contributor:
    fullname: Yu
– start-page: 2546
  year: 2011
  end-page: 2554
  ident: b0015
  article-title: Algorithms for hyper-parameter optimization
  publication-title: Advances in Neural Information Processing Systems
  contributor:
    fullname: Kégl
– volume: 102
  start-page: 207
  year: 2018
  end-page: 217
  ident: b0105
  article-title: Predicting mortgage default using convolutional neural networks
  publication-title: Expert Systems with Applications
  contributor:
    fullname: Sjursen
– volume: 70
  start-page: 489
  year: 2006
  end-page: 501
  ident: b0075
  article-title: Extreme learning machine: theory and applications
  publication-title: Neurocomputing
  contributor:
    fullname: Siew
– volume: 521
  start-page: 436
  issue: 7553
  year: 2015
  ident: 10.1016/j.eswa.2020.114185_b0110
  article-title: Deep learning
  publication-title: Nature
  doi: 10.1038/nature14539
  contributor:
    fullname: LeCun
– volume: 13
  start-page: 21
  issue: 1
  year: 1967
  ident: 10.1016/j.eswa.2020.114185_b0050
  article-title: Nearest neighbor pattern classification
  publication-title: IEEE Transactions on Information Theory
  doi: 10.1109/TIT.1967.1053964
  contributor:
    fullname: Cover
– ident: 10.1016/j.eswa.2020.114185_b0055
– ident: 10.1016/j.eswa.2020.114185_b0080
– ident: 10.1016/j.eswa.2020.114185_b0145
– volume: 115
  start-page: 211
  issue: 3
  year: 2015
  ident: 10.1016/j.eswa.2020.114185_b0130
  article-title: ImageNet large scale visual recognition challenge
  publication-title: International Journal of Computer Vision
  doi: 10.1007/s11263-015-0816-y
  contributor:
    fullname: Russakovsky
– volume: 45
  start-page: 267
  issue: 3
  year: 2003
  ident: 10.1016/j.eswa.2020.114185_b0180
  article-title: The Elements of Statistical Learning
  publication-title: Technometrics
  doi: 10.1198/tech.2003.s770
  contributor:
    fullname: Ziegel
– volume: 102
  start-page: 207
  year: 2018
  ident: 10.1016/j.eswa.2020.114185_b0105
  article-title: Predicting mortgage default using convolutional neural networks
  publication-title: Expert Systems with Applications
  doi: 10.1016/j.eswa.2018.02.029
  contributor:
    fullname: Kvamme
– ident: 10.1016/j.eswa.2020.114185_b0095
– volume: 60
  start-page: 84
  issue: 6
  year: 2017
  ident: 10.1016/j.eswa.2020.114185_b0100
  article-title: ImageNet classification with deep convolutional neural networks
  publication-title: Advances in Neural Information Processing Systems
  contributor:
    fullname: Krizhevsky
– volume: 159
  start-page: 113408
  year: 2020
  ident: 10.1016/j.eswa.2020.114185_b0060
  article-title: An efficient convolutional neural network for coronary heart disease prediction
  publication-title: Expert Systems with Applications
  doi: 10.1016/j.eswa.2020.113408
  contributor:
    fullname: Dutta
– year: 2017
  ident: 10.1016/j.eswa.2020.114185_b0030
  article-title: Classification and regression trees
  publication-title: Classification and Regression Trees
  doi: 10.1201/9781315139470-8
  contributor:
    fullname: Breiman
– volume: 11
  start-page: 354
  year: 1998
  ident: 10.1016/j.eswa.2020.114185_b0070
  article-title: An introduction to structural equation modeling
  publication-title: Family Science Review
  contributor:
    fullname: Hox
– ident: 10.1016/j.eswa.2020.114185_b0035
  doi: 10.1145/2939672.2939785
– volume: 12
  start-page: 145
  issue: 1
  year: 1999
  ident: 10.1016/j.eswa.2020.114185_b0120
  article-title: On the momentum term in gradient descent learning algorithms
  publication-title: Neural Networks
  doi: 10.1016/S0893-6080(98)00116-6
  contributor:
    fullname: Qian
– volume: 70
  start-page: 489
  issue: 1-3
  year: 2006
  ident: 10.1016/j.eswa.2020.114185_b0075
  article-title: Extreme learning machine: theory and applications
  publication-title: Neurocomputing
  doi: 10.1016/j.neucom.2005.12.126
  contributor:
    fullname: Huang
– volume: 1
  start-page: 81
  issue: 1
  year: 1986
  ident: 10.1016/j.eswa.2020.114185_b0125
  article-title: Induction of decision trees
  publication-title: Mach Learn
  doi: 10.1007/BF00116251
  contributor:
    fullname: Quinlan
– volume: 22
  start-page: 39
  issue: 1
  year: 1996
  ident: 10.1016/j.eswa.2020.114185_b0005
  article-title: A Maximum Entropy Approach to Natural Language Processing
  publication-title: Computational Linguistics
  contributor:
    fullname: Berger
– ident: 10.1016/j.eswa.2020.114185_b0140
– ident: 10.1016/j.eswa.2020.114185_b0155
  doi: 10.1109/CVPR.2014.220
– ident: 10.1016/j.eswa.2020.114185_b0065
  doi: 10.1109/CVPR.2016.90
– ident: 10.1016/j.eswa.2020.114185_b0090
  doi: 10.3115/v1/D14-1181
– start-page: 2546
  year: 2011
  ident: 10.1016/j.eswa.2020.114185_b0015
  article-title: Algorithms for hyper-parameter optimization
  publication-title: Advances in Neural Information Processing Systems
  contributor:
    fullname: Bergstra
– volume: 2
  start-page: 1401
  year: 1999
  ident: 10.1016/j.eswa.2020.114185_b0135
  article-title: A Brief Introduction to Boosting
  publication-title: IJCAI International Joint Conference on Artificial Intelligence
  contributor:
    fullname: Schapire
– ident: 10.1016/j.eswa.2020.114185_b0040
  doi: 10.1145/1390156.1390177
– volume: 20
  start-page: 273
  issue: 3
  year: 1995
  ident: 10.1016/j.eswa.2020.114185_b0045
  article-title: Support-vector networks
  publication-title: Mach Learn
  doi: 10.1007/BF00994018
  contributor:
    fullname: Cortes
– ident: 10.1016/j.eswa.2020.114185_b0085
– ident: 10.1016/j.eswa.2020.114185_b0150
  doi: 10.1109/ICCV.2011.6126456
– volume: 13
  start-page: 281
  issue: 1
  year: 2012
  ident: 10.1016/j.eswa.2020.114185_b0010
  article-title: Random search for hyper-parameter optimization
  publication-title: The Journal of Machine Learning Research
  contributor:
    fullname: Bergstra
– year: 2008
  ident: 10.1016/j.eswa.2020.114185_b0165
  article-title: Visualizing data using t-SNE
  publication-title: Journal of Machine Learning Research
  contributor:
    fullname: Van Der Maaten
– volume: 323
  start-page: 37
  year: 2019
  ident: 10.1016/j.eswa.2020.114185_b0175
  article-title: Recent advances in convolutional neural network acceleration
  publication-title: Neurocomputing
  doi: 10.1016/j.neucom.2018.09.038
  contributor:
    fullname: Zhang
– year: 1998
  ident: 10.1016/j.eswa.2020.114185_b0115
  article-title: The MNIST database of handwritten digits
  publication-title: The Courant Institute of Mathematical Sciences.
  contributor:
    fullname: LeCun
– ident: 10.1016/j.eswa.2020.114185_b0170
  doi: 10.1109/CVPR.2018.00813
– volume: 45
  start-page: 5
  issue: 1
  year: 2001
  ident: 10.1016/j.eswa.2020.114185_b0025
  article-title: Random forests
  publication-title: Machine Learning
  doi: 10.1023/A:1010933404324
  contributor:
    fullname: Breiman
– ident: 10.1016/j.eswa.2020.114185_b0020
  doi: 10.1088/1749-4699/8/1/014008
SSID ssj0017007
Score 2.4269865
Snippet •Propose a feature grid-based CNN model, FGCN, on traditional data.•Propose methods of converting instance with form of 1-d vector to feature grid.•The...
Convolutional neural networks (ConvNets) have been applied to various types of data, including image, text, and speech, but not to traditional data. In this...
SourceID proquest
crossref
elsevier
SourceType Aggregation Database
Publisher
StartPage 114185
SubjectTerms Artificial neural networks
Classification
Convolutional neural network
Data conversion
Neural networks
Traditional data
Title Application of convolutional neural network to traditional data
URI https://dx.doi.org/10.1016/j.eswa.2020.114185
https://www.proquest.com/docview/2492315819/abstract/
Volume 168
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwnV07T8MwED71sbDwRhRKlYENmcbPJBOqKqoCogtU6mbFdiyVoa0giI3fjp04IBBiYIqU5KLos30P3Xd3AOcp0cbFyRnydZ6IEaxRShlHsciIMbHVyvqM7v1MTOfsdsEXLRg3tTCeVhl0f63TK20d7gwDmsPNcjl8cM6BM4cutPNZAueotKHrzBFjHeiObu6ms89kQhLXVdPufeQFQu1MTfMqXt58-yFSdc3FfqTy7_bph6auzM9kF7aD3xiN6l_bg1ax2oedZiZDFI7oAVyNvjLS0dpGnlUedpcT990rq0vF_Y7KdVQ-52YZnnq26CHMJ9eP4ykKQxKQphkrkc7SOMfc6pgUmCvOsE0ts9a4wCdLMm4JKRinllCTKJrlAgtFcUE0Fppwi-kRdFbrVXEMkSYmETgXRrkYytl2lXjyqRa5SmNFuenBRQON3NS9MGRDEnuSHkjpgZQ1kD3gDXry24pKp6z_lOs3UMtwnl4kqfrIcee-nPzzs6ewRTwdxbdp5H3olM-vxZnzJ0o1gPblOx6EXfMBKXjH_w
link.rule.ids 315,786,790,4521,24144,27957,27958,45620,45714
linkProvider Elsevier
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwnV1LS8NAEB5qe9CLb7FaNQdvEpp95XGSUiypfVxsobclu5uFemhLG_Hvu5tsFEU8eAokmRC-3Z0H880MwH2MpTJxcuLbOk-fYiT9mFDmB2GClQq0FNpmdCfTMJ3T5wVbNKBf18JYWqXT_ZVOL7W1u9N1aHY3y2X3xTgHxhya0M5mCYyjsgctyiJEm9DqDUfp9DOZEAVV1bR537cCrnamonnlu3fbfgiXXXORHan8u336oalL8zM4hkPnN3q96tdOoJGvTuGonsnguSN6Bo-9r4y0t9aeZZW73WXEbffK8lJyv71i7RXbTC3dU8sWPYf54GnWT303JMGXJKGFL5M4yBDTMsA5YoJRpGNNtVYm8EmihGmMc8qIxkRFgiRZiEJBUI4lCiVmGpELaK7Wq_wSPIlVFKIsVMLEUMa2i8iST2WYiTgQhKk2PNTQ8E3VC4PXJLFXboHkFkheAdkGVqPHv60oN8r6T7lODTV352nHcdlHjhn35eqfn72D_XQ2GfPxcDq6hgNsqSm2ZSPrQLPYvuU3xrcoxK3bOx8ATsnx
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Application+of+convolutional+neural+network+to+traditional+data&rft.jtitle=Expert+systems+with+applications&rft.au=Zhang%2C+Xiaohang&rft.au=Wu%2C+Fengmin&rft.au=Li%2C+Zhengren&rft.date=2021-04-15&rft.pub=Elsevier+BV&rft.issn=0957-4174&rft.eissn=1873-6793&rft.volume=168&rft.spage=1&rft_id=info:doi/10.1016%2Fj.eswa.2020.114185&rft.externalDBID=NO_FULL_TEXT
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=0957-4174&client=summon
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=0957-4174&client=summon
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=0957-4174&client=summon