A new method using the convolutional neural network with compressive sensing for fabric defect classification based on small sample sizes

The convolutional neural network (CNN) has recently achieved great breakthroughs in many computer vision tasks. However, its application in fabric texture defects classification has not been thoroughly researched. To this end, this paper carries out a research on its application based on the CNN mod...

Full description

Saved in:
Bibliographic Details
Published inTextile research journal Vol. 89; no. 17; pp. 3539 - 3555
Main Authors Wei, Bing, Hao, Kuangrong, Tang, Xue-song, Ding, Yongsheng
Format Journal Article
LanguageEnglish
Published London, England SAGE Publications 01.09.2019
Sage Publications Ltd
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:The convolutional neural network (CNN) has recently achieved great breakthroughs in many computer vision tasks. However, its application in fabric texture defects classification has not been thoroughly researched. To this end, this paper carries out a research on its application based on the CNN model. Meanwhile, since the CNN cannot achieve good classification accuracy in small sample sizes, a new method combining compressive sensing and the convolutional neural network (CS-CNN) is proposed. Specifically, this paper uses the compressive sampling theorem to compress and augment the data in small sample sizes; then the CNN can be employed to classify the data features directly from compressive sampling; finally, we use the test data to verify the classification performance of the method. The explanatory experimental results demonstrate that, in comparison with the state-of-the-art methods for running time, our CS-CNN approach can effectively improve the classification accuracy in fabric defect samples, even with a small number of defect samples.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
content type line 23
ISSN:0040-5175
1746-7748
1746-7748
DOI:10.1177/0040517518813656