FMatchNet algorithm for fast clothing matching
- Vol. 24, Issue 6, Pages: 979-986(2019)
Received:11 September 2018,
Revised:2019-1-13,
Published:16 June 2019
DOI: 10.11834/jig.180538
移动端阅览

浏览全部资源
扫码关注微信
Received:11 September 2018,
Revised:2019-1-13,
Published:16 June 2019
移动端阅览
目的
2
针对现有服装搭配系统中,提取服装图像深度特征进行搭配所需时间过长的问题,提出了一种新的FMatchNet网络提取哈希特征进行服装快速搭配的方法。
方法
2
首先采用快速区域卷积神经网络(Faster-RCNN)方法检测出图像中的服装,用此服装进行搭配可以最大限度地保留服装信息并消除背景信息的干扰。然后用深度卷积神经网络提取服装的深度特征并产生服装的哈希码,采用查询扩展的方法完成服装搭配。模型采用Siamese网络的训练方法使哈希码尽可能保留服装图像的语义信息。另外,由于目前国际上缺少大型时尚服装数据库,本文扩建了一个细粒度标注的时尚服装数据库。
结果
2
在FClothes数据库上验证本文方法并与目前流行的方法进行对比,本文方法在哈希长度为16时,上、下服装搭配方面的准确度达到了50.81%,搭配速度相对于基本准线算法提高了近3倍。
结论
2
针对大规模服装搭配问题,提出一种新的FMatchNet网络提取特征进行服装快速搭配的方法,提高了服装搭配的精度和速度,适用于日常服装搭配。
Objective
2
With the development of artificial intelligence and online shopping
clothing matching based on clothing images is crucial in helping merchants promote sales. An increasing number of young consumers are inclined to buy clothing online
but existing research mainly focuses on clothing search
clothing recommendations
and fashion trends. Quickly
accurately
and effectively matching the right clothing to the clothing that the user has already purchased remains a challenging task. With the development of the economy and the improvement of material level
the clothing style and number are increasing. Therefore
clothing matching among a large number of garments is vital. Aiming at the problem that the existing clothing matching framework is used in fashion clothing matching and the depth feature of clothing image extraction requires a large time overhead
this study proposes a new FMatchNet network for extracting hash features for a fast clothing matching.
Method
2
Deep learning is an important development in the field of machine learning and artificial intelligence. At present
deep convolutional neural networks have become one of the most effective means of extracting image features. The early extraction feature method is based on artificial extraction features
such as scale-invariant feature transform
speeded-up robust features
and histogram of oriented gradient. The features extracted by deep neural networks are more accurate than traditional features. By contrast
the use of binary Hashing codes for image features is an effective approach for reducing overhead and increasing computational speed. The core of the clothing mix is the description of the clothing image content. To match the clothing efficiently
the content of the clothing image must be described
and the basic idea is to express the clothing image as a feature vector. In general
the more closely matched the clothing images
the smaller the distance between their feature vectors. Recent studies in many image fields have begun to explore methods for generating Hashing codes on the basis of features extracted by deep networks. This research also applies this idea to study the clothing image representation method combined with deep learning and Hashing code. This study proposes a fast
accurate
and effective clothing matching network
namely
FMatchNet. The faster regional convolutional neural network (Faster-RCNN) method is adopted to detect the clothing area in the image. The clothing area can be used to maximize the original clothing information and eliminate the interference of image background information. Then
the image of the clothing area extracts the depth feature of the garment and the Hashing code of the garment through a two-way deep convolutional neural network. Finally
clothing matching is completed by using the query expansion method. The model applies the Siamese network training method to extract the depth features of the clothing image and the Hashing code extracted by this method
which can preserve the semantic information of the clothing image as much as possible. The Hashing code is used to select the candidate set of clothing matching
and then depth feature is used to rank the clothing matching in the selected clothing matching candidate set. In addition
given the lack of large-scale fashion clothing database in the world
a fine-grained fashion apparel database has been expanded in this paper. The expanded FClothes clothing dataset and dataset images mainly come from the Weibo website
which contains a large number of popular people and high-resolution clothing pictures that meet fashion demands. Finally
the algorithm is experimentally verified on the expanded fine-grained fashion clothing database.
Result
2
The method used in this study was verified on the expanded FClothes database and compared with current popular methods. This study compares the 8 bits
16 bits
and 32 bits Hashing codes
and experimental results show that as the length of the Hashing code increases
the precision of the clothing mix increases. However
the time consumption also increases. When the lengths of the Hashing code are 16 and 32 bits
the matching accuracy of the upper and lower garments is higher than that of the baseline. When the length of the Hashing code is 16 bits
the matching accuracy of the method used in this paper is 50.81% in the upper and lower clothing
and the matching speed is nearly three times higher than that of the basic alignment algorithm. The basic accuracy of the comparison comes from the "Learning visual clothing style with heterogeneous dyadic co-occurrences" of the International Conference on Computer Vision 2015. From this point of view
the accuracy and time of the upper and lower clothing combinations of the algorithm are better than those of the current cutting-edge methods.
Conclusion
2
In view of the problem of large-scale clothing matching
this study proposes a new FMatchNet network extraction feature
which improves the precision and speed of clothing matching and is suitable for daily clothing matching.
Lindeberg T. Scale invariant feature transform[J]. Scholarpedia, 2012, 7(5):#10491.[DOI:10.4249/scholarpedia.10491]
Huang X, Ling Z G, Li X X.Discriminative deep feature learning method by fusing linear discriminant analysis for image recognition[J]. Journal of Image and Graphics, 2018, 23(4):510-518.
黄旭, 凌志刚, 李绣心.融合判别式深度特征学习的图像识别算法[J].中国图象图形学报, 2018, 23(4):510-518. [DOI:10.11834/jig.170336]
Ren S Q, He K M, Girshick R, et al. Faster R-CNN: towards real-time object detection with region proposal networks[C]//Proceedings of the 28th International Conference on Neural Information Processing Systems. Montreal, Canada: ACM, 2015: 91-99. https://arxiv.org/abs/1506.01497 .
Liu Y J, Gao Y B, Feng S H, et al. Weather-to-garment: weather-oriented clothing recommendation[C]//Proceedings of 2017 IEEE International Conference on Multimedia and Expo. Hong Kong, China: IEEE, 2017: 181-186.[ DOI: 10.1109/ICME.2017.8019476 http://dx.doi.org/10.1109/ICME.2017.8019476 ]
Liu S, Song Z, Liu G C, et al. Street-to-shop: cross-scenario clothing retrieval via parts alignment and auxiliary set[C]//Proceedings of 2012 IEEE Conference on Computer Vision and Pattern Recognition. Providence, RI, USA: IEEE, 2012: 3330-3337.[ DOI: 10.1109/CVPR.2012.6248071 http://dx.doi.org/10.1109/CVPR.2012.6248071 ]
Liu S, Feng J S, Zhang T Z, et al. Hi, magic closet, tell me what to wear![C]//Proceedings of the 20th ACM International Conference on Multimedia. Nara, Japan: ACM, 2012: 619-628.[ DOI: 10.1145/2393347.2393433 http://dx.doi.org/10.1145/2393347.2393433 ]
Li Y C, Cao L L, Zhu J, et al. Mining fashion outfit composition using an end-to-end deep learning approach on set data[J]. IEEE Transactions on Multimedia, 2017, 19(8):1946-1955.[DOI:10.1109/TMM.2017.2690144]
Felzenszwalb P, McAllester D, Ramanan D. A discriminatively trained, multiscale, deformable part model[C]//Proceedings of 2008 IEEEConference on Computer Vision and Pattern Recognition. Anchorage, AK, USA: IEEE, 2008: 1-8.[ DOI: 10.1109/CVPR.2008.4587597 http://dx.doi.org/10.1109/CVPR.2008.4587597 ]
Zhang Z H, Zhou C L, Liang Y.An optimized clothing classification algorithm based on residual convolutional neural network[J]. Computer Engineering and Science, 2018.
张振焕, 周彩兰, 梁媛.基于残差的优化卷积神经网络服装分类算法[J].计算机工程与科学, 2018.]
Yang T Q, Huang S X. Application of improved convolution neural network in classification and recommendation[J]. Application Research of Computers, 2018, 35(4):974-977.
杨天祺, 黄双喜.改进卷积神经网络在分类与推荐中的实例应用[J].计算机应用研究, 2018, 35(4):974-977].
Yuan W F, Guo J M, Su Z, et al. Clothing retrieval by deep muti-label parsing and Hashing[J].Journal of Image and Graphics, 2019, 24(2):159-169.
原尉峰, 郭佳明, 苏卓, 等.结合深度多标签解析的哈希服装检索[J].中国图象图形学报, 2019, 24(2):159-169. [DOI:10.11834/jig.180361]
Veit A, Kovacs B, Bell S, et al. Learning visual clothing style with heterogeneous dyadic co-occurrences[C]//Proceedings of 2015 IEEE International Conference on Computer Vision. Santiago, Chile: IEEE, 2015: 4642-4650.[ DOI: 10.1109/ICCV.2015.527 http://dx.doi.org/10.1109/ICCV.2015.527 ]
Song X M, Feng F L, Liu J H, et al. NeuroStylist: neural compatibility modeling for clothing matching[C]//Proceedings of the 25th ACM International Conference on Multimedia. Mountain View, California, USA: ACM, 2017: 753-761.[ DOI: 10.1145/3123266.3123314 http://dx.doi.org/10.1145/3123266.3123314 ]
Gionis A, Indyk P, Motwani R. Similarity search in high dimensions via Hashing[C]//Proceedings of the 25th International Conference on Very Large Data Bases. Edinburgh, Scotland, UK: ACM, 1999: 518-529. https://wenku.baidu.com/view/8fa10409bb68a98271fefa64.html .
Weiss Y, Torralba A, Fergus R. Spectral Hashing[C]//Proceedings of the 21st International Conference on Neural Information Processing Systems. Vancouver, British Columbia, Canada: ACM, 2008: 1753-1760. http://people.csail.mit.edu/torralba/publications/spectralhashing.pdf .
Gong Y C, Lazebnik S. Iterative quantization: a procrustean approach to learning binary codes[C]//Proceedings of 2011 IEEE Conference on Computer Vision and Pattern Recognition. Colorado Springs, CO, USA: IEEE, 2011: 817-824.[ DOI: 10.1109/CVPR.2011.5995432 http://dx.doi.org/10.1109/CVPR.2011.5995432 ]
Norouzi M, Fleet D J. Minimal loss Hashing for compact binary codes[C]//Proceedings of the 28th International Conference on International Conference on Machine Learning. Bellevue, Washington, USA: ACM, 2011: 353-360. http://www.cs.toronto.edu/~norouzi/research/papers/min_loss_hash.pdf .
Liu W, Wang J, Ji R R, et al. Supervised Hashing with kernels[C]//Proceedings of 2012 IEEE Conference on Computer Vision and Pattern Recognition. Providence, RI, USA: IEEE, 2012: 2074-2081.[ DOI: 10.1109/CVPR.2012.6247912 http://dx.doi.org/10.1109/CVPR.2012.6247912 ]
Kulis B, Darrell T. Learning to Hash with binary reconstructive embeddings[C]//Proceedings of the 22nd International Conference on Neural Information Processing Systems. Vancouver, British Columbia, Canada: ACM, 2009: 1042-1050. https://www2.eecs.berkeley.edu/Pubs/TechRpts/2009/EECS-2009-101.pdf .
Lai H J, Pan Y, Liu Y, et al. Simultaneous feature learning and Hash coding with deep neural networks[C]//Proceedings of 2015 IEEE Conference on Computer Vision and Pattern Recognition. Boston, MA, USA: IEEE, 2015: 3270-3278.[ DOI: 10.1109/CVPR.2015.7298947 http://dx.doi.org/10.1109/CVPR.2015.7298947 ]
Fang Z W, Liu J, Wang Y H, et al. Object-aware deep network for commodity image retrieval[C]//Proceedings of 2016 ACM on International Conference on Multimedia Retrieval. New York, USA: ACM, 2016: 405-408.[ DOI: 10.1145/2911996.2912027 http://dx.doi.org/10.1145/2911996.2912027 ]
Lin K, Yang H F, Hsiao J H, et al. Deep learning of binary Hash codes for fast image retrieval[C]//Proceedings of 2015 IEEE Conference on Computer Vision and Pattern Recognition Workshops. Boston, MA, USA: IEEE, 2015: 27-35.[ DOI: 10.1109/CVPRW.2015.7301269 http://dx.doi.org/10.1109/CVPRW.2015.7301269 ]
Melekhov I, Kannala J, Rahtu E. Siamese network features for image matching[C]//Proceedings of 2016 International Conference on Pattern Recognition.Cancun, Mexico: IEEE, 2016: 378-383.[ DOI: 10.1109/ICPR.2016.7899663 http://dx.doi.org/10.1109/ICPR.2016.7899663 ]
Chum O, Philbin J, Sivic J, et al. Total recall: automatic query expansion with a generative feature model for object retrieval[C]//Proceedings of 2007 IEEE International Conference on Computer Vision. Rio de Janeiro, Brazil: IEEE, 2007: 1-8.[ DOI: 10.1109/ICCV.2007.4408891 http://dx.doi.org/10.1109/ICCV.2007.4408891 ]
Kiapour M H, Han X F, Lazebnik S, et al. Where to buy it: matching street clothing photos in online shops[C]//Proceedings of 2015 IEEE International Conference on Computer Vision. Santiago, Chile: IEEE, 2015: 3343-3351.[ DOI: 10.1109/ICCV.2015.382 http://dx.doi.org/10.1109/ICCV.2015.382 ]
Yang Y, Ramanan D. Articulated pose estimation with flexible mixtures-of-parts[C]//Proceedings of 2011 IEEE Conference on Computer Vision and Pattern Recognition. Colorado Springs, CO, USA: IEEE, 2011: 1385-1392.[ DOI: 10.1109/CVPR.2011.5995741 http://dx.doi.org/10.1109/CVPR.2011.5995741 ]
Liang X D, Lin L, Yang W, et al. Clothes co-parsing via joint image segmentation and labeling with application to clothing retrieval[J]. IEEE Transactions on Multimedia, 2016, 18(6):1175-1186.[DOI:10.1109/TMM.2016.2542983]
Yamaguchi K, Kiapour M H, Ortiz L E, et al. Parsing clothing in fashion photographs[C]//Proceedings of 2012 IEEE Conference on Computer Vision and Pattern Recognition. Providence, RI, USA: IEEE, 2012: 3570-3577.[ DOI: 10.1109/CVPR.2012.6248101 http://dx.doi.org/10.1109/CVPR.2012.6248101 ]
相关文章
相关作者
相关机构
京公网安备11010802024621