1,278
Views
15
CrossRef citations to date
0
Altmetric
Research Article

Image Aesthetics Assessment Based on Multi-stream CNN Architecture and Saliency Features

, &

References

  • Abrar, H. A., W. Gang, L. Jiwen, and J. Kui. 2015. Multi-task CNN model for attribute prediction. IEEE Transactions on Multi-Scale 17 (11):1949–59. doi:https://doi.org/10.1109/TMM.2015.2477680.
  • Borji, A., and L. Itti. 2015. CAT2000: A large scale fixation dataset for boosting saliency research, Proc. of IEEE Int. Conf. on Comp. Vision and Pattern Recog., Workshop on “Future of Datasets. arXiv Preprint arXiv 1505.03581.
  • Chenarlogh, V. A., and F. Razzazi. 2019. “Multi-stream 3D CNN structure for human action recognition trained by limited data”. IET Computer Vision 13 (4):338–44. doi:https://doi.org/10.1049/iet-cvi.2018.5088.
  • Chollet, F. 2016. Xception: Deep learning with depthwise separable convolutions. Proc. of 2017 IEEE Conf. on Comp. Vision and Pattern Recog., 1800–07.
  • Coe, K. 1992. Art: The replicable unit - An inquiry into the possible origin of art as a social behavior”. Journal of Social and Evolutionary Systems 15 (2):217–34. doi:https://doi.org/10.1016/1061-7361(92)90005-X.
  • Csurka, G., C. Dance, L. Fan, J. Willamowski, and C. Bray. 2004. Visual categorization with bags of keypoints. Proc. European Conference on Computer Vision Workshop on Statistical Learning in Computer Vision, 59–74.
  • Datta, R., D. Joshi, J. Li, and J. Z. Wang. 2006. Studying aesthetics in photographic images using a computational approach. Proc. of the 9th European Conference on Computer Vision, 288–301.
  • Deng, J., W. Dong, R. Socher, L. J. Li, K. Li, and L. Fei-Fei. 2009. ImageNet: A large-scale hierarchical image database. Proceedings of the IEEE Computer Conference on Computer Vision and Pattern Recognition 248–55.
  • Dhar, S., V. Ordonez, and T. L. Berg. 2011. High level describable attributes for predicting aesthetics and interestingness. Proc. of 2011 IEEE Conf. on Comp. Vision and Pattern Recog, 1657–64.
  • Huang, X., C. Shen, X. Boix, and Q. Zhao. 2015. SALICON: Reducing the semantic gap in saliency prediction by adapting deep neural networks. Proceedings of IEEE International Conference on Computer Vision 262–70.
  • Itti, L., C. Koch, and E. Niebur. 1998. “A model of saliency-based visual attention for rapid scene analysis”. IEEE Transactions on Pattern Analysis and Machine Intelligence 20 (11):1254–59. doi:https://doi.org/10.1109/34.730558.
  • Ke, Y., X. Tang, and F. Jing. 2006. The design of high-level features for photo quality assessment. Proceedings of IEEE Conference on Computer Vision and Pattern Recognition 419–26.
  • Kimura, A., R. Yonetani, and T. Hirayama. 2013. Computational models of human visual attention and their implementations: A survey. IEICE Transactions on Information and Systems 96 (3):562––578.
  • Koch, C., and S. Ullman. 1985. Shifts in selective visual attention: Towards the underlying neural circuitry. Human Neurobiology 4 (4):219––227.
  • Kong, S., X. Shen, Z. Lin, R. Mech, and C. Fowlkes. 2016. Photo aesthetics ranking network with attributes and content adaptation. Proceedings of European Conference on Computer Vision 662–79.
  • LeCun, Y., Y. Bengio, and G. E. Hinton. 2015. Deep Learning. Nature 521:436–44.
  • Lind, R. W. 1980. Attention and the aesthetics object. Journal of Aesthetics and Art Criticism 39 (2):131–42. doi:https://doi.org/10.2307/429807.
  • Lu, X., Z. Lin, H. Jin, J. Yang, and J. Z. Wang. 2015a. Rating image aesthetics using deep learning. IEEE Transactions on Multimedia 17 (11):2021–34. doi:https://doi.org/10.1109/TMM.2015.2477040.
  • Lu, X., Z. Lin, X. Shen, R. Mech, and J. Z. Wang. 2015b. Deep multi-patch aggregation network for image style, aesthetics, and quality estimation. Proceedings of IEEE International Conference on Computer Vision 990––998.
  • Luo, W., X. Wang, and X. Tang. 2011. Content-based photo quality assessment. Proceedings of the IEEE International Conference on Computer Vision 2206–13.
  • Ma, S., J. Liu, and C. W. Chen. 2017. A-lamp: Adaptive layout-aware multipatch deep convolutional neural network for photo aesthetic assessment. Proceedings of the IEEE Computer Conference on Computer Vision and Pattern Recognition 722–31.
  • Mai, L., H. Jin, and F. Liu. 2016. Composition-preserving deep photo aesthetics assessment. Proceedings of the IEEE Computer Conference on Computer Vision and Pattern Recognition 2016:pp. 497–506.
  • Marchesotti, L., F. Perronnin, D. Larlus, and G. Csurka. 2011. Assessing the aesthetic quality of photographs using generic image descriptors. Proc. of IEEE Int. Conf. on Comp. Vision, 1784–91.
  • Murray, N., L. Marchesotti, and F. Perronnin. 2012. AVA: A large-scale database for aesthetic visual analysis. Proceedings of the IEEE Computer Conference on Computer Vision and Pattern Recognition 2408–15.
  • Nishiyama, M., T. Okabe, I. Sato, and Y. Sato. 2011. Aesthetic quality classification of photographs based on color harmony. Proceedings of IEEE Conference on Computer Vision and Pattern Recognition 33–40.
  • Oliva, A., and A. Torralba. 2001. Modeling the shape of the scene: A holistic representation of the spatial envelope. International Journal of Computer Vision 42 (3):145––175. doi:https://doi.org/10.1023/A:1011139631724.
  • Omori, F., H. Takimoto, H. Yamauchi, A. Kanagawa, T. Iwasaki, and M. Ombe. 2019. Aesthetic quality evaluation using convolutional neural network. Asia-Pacific Journal of Industrial Management 8 (1):71–77.
  • Perronnin, F., and C. Dance. 2007. Fisher kernels on visual vocabularies for image categorization. Proceedings of the IEEE Computer Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 1–8.
  • Sulfayanti, F. S., H. Takimoto, S. Sato, H. Yamauchi, A. Kanagawa, and A. Lawi. 2019. Food constituent estimation for lifestyle disease prevention by multi-task CNN. Applied Artificial Intelligence 33 (8):732–46. doi:https://doi.org/10.1080/08839514.2019.1602318.
  • Takimoto, H., S. Katsumata, S. F. Situju, A. Kanagawa, and A. Lawi. 2018. Visual saliency estimation based on multi-task CNN, Proc. of The Fourteenth International Conference on Industrial Management (ICIM2018), Hangzhou, China 639––643.
  • Talebi, H., and P. Milanfar. 2018. NIMA: Neural image assessment. IEEE Transactions on Image Processing 27 (8):3998–4011. doi:https://doi.org/10.1109/TIP.2018.2831899.
  • Treisman, A., and G. Gelade. 1980. “A feature-integration theory of attention”. Cognitive Psychology 12 (1):97––136. doi:https://doi.org/10.1016/0010-0285(80)90005-5.
  • Tu, Z., W. Xie, Q. Qin, R. Poppe, R. Veltkamp, B. Li, and J. Yuan. 2018. Multi-stream CNN: Learning representations based on human-related regions for action recognition. Pattern Recognition 79:32–43. doi:https://doi.org/10.1016/j.patcog.2018.01.020.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.