|
{ |
|
"paper_id": "2020", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T02:09:33.383501Z" |
|
}, |
|
"title": "Toward the Automatic Retrieval and Annotation of Outsider Art images: A Preliminary Statement", |
|
"authors": [ |
|
{ |
|
"first": "John", |
|
"middle": [], |
|
"last": "Roberto", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "INSIGHT Centre Dublin City University", |
|
"location": { |
|
"settlement": "Glasnevin, Dublin 9", |
|
"country": "Ireland" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Diego", |
|
"middle": [], |
|
"last": "Ortego", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "INSIGHT Centre Dublin City University", |
|
"location": { |
|
"settlement": "Glasnevin, Dublin 9", |
|
"country": "Ireland" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Brian", |
|
"middle": [], |
|
"last": "Davis", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "INSIGHT Centre Dublin City University", |
|
"location": { |
|
"settlement": "Glasnevin, Dublin 9", |
|
"country": "Ireland" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "The aim of this position paper is to establish an initial approach to the automatic classification of digital images about the Outsider Art style of painting. Specifically, we explore whether is it possible to classify non-traditional artistic styles by using the same features that are used for classifying traditional styles? Our research question is motivated by two facts. First, art historians state that non-traditional styles are influenced by factors \"outside\" of the world of art. Second, some studies have shown that several artistic styles confound certain classification techniques. Following current approaches to style prediction, this paper utilises Deep Learning methods to encode image features. Our preliminary experiments have provided motivation to think that, as is the case with traditional styles, Outsider Art can be computationally modelled with objective means by using training datasets and CNN models. Nevertheless, our results are not conclusive due to the lack of a large available dataset on Outsider Art. Therefore, at the end of the paper, we have mapped future lines of action, which include the compilation of a large dataset of Outsider Art images and the creation of an ontology of Outsider Art. This research forms part of a wider project called \"Semantic Analysis of Text Corpora in the Outsider Art Domain\".", |
|
"pdf_parse": { |
|
"paper_id": "2020", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "The aim of this position paper is to establish an initial approach to the automatic classification of digital images about the Outsider Art style of painting. Specifically, we explore whether is it possible to classify non-traditional artistic styles by using the same features that are used for classifying traditional styles? Our research question is motivated by two facts. First, art historians state that non-traditional styles are influenced by factors \"outside\" of the world of art. Second, some studies have shown that several artistic styles confound certain classification techniques. Following current approaches to style prediction, this paper utilises Deep Learning methods to encode image features. Our preliminary experiments have provided motivation to think that, as is the case with traditional styles, Outsider Art can be computationally modelled with objective means by using training datasets and CNN models. Nevertheless, our results are not conclusive due to the lack of a large available dataset on Outsider Art. Therefore, at the end of the paper, we have mapped future lines of action, which include the compilation of a large dataset of Outsider Art images and the creation of an ontology of Outsider Art. This research forms part of a wider project called \"Semantic Analysis of Text Corpora in the Outsider Art Domain\".", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "This paper is about the computational analysis of visual aesthetics. We focus our attention on Outsider Art, which is considered by some as the \"unsightly style\". At present, aesthetics constitutes a field of interest for scientists working in Artificial Intelligence, particularly in the context of paintings. Five of the main tasks in this field are: the prediction of ratings, the detection of forgery in paintings, artist identification, genre recognition and style prediction. First, the prediction of ratings (Talebi and Milanfar, 2017) captures the technical and semantic level characteristics associated with emotions and beauty in images in order to categorize images in two classes: low and high quality. Second, the detection of forgery in paintings (Mane, 2017) assumes that an artist's brushwork is characterized by signature features that can be detected automatically. Third, \"artist identification is the task of identifying the artist of a painting given no other information about it\" (Viswanathan, 2017) . Fourth, genre recognition in paintings (Agarwal et al., 2015) focuses on classifying works of art according to the (type of) scene that is depicted by the artist. Finally, style prediction uses both low-level and semantic features in order to group paintings according to their shared properties. Several studies on style prediction will be commented on this paper. Recently, deep learning methods have been growing in popularity for style classification because they can achieve state-of-the-art performance in this field. For example, Deep Convolutional Neural Networks models such as AlexNet, VGGNet and ResNet have been applied to the classification of traditional painting styles with varying success thanks to the existence of large scale datasets of digital paintings. However, to the best of our knowledge, there are no attempts to classify, retrieve and annotate the Outsider Art style.", |
|
"cite_spans": [ |
|
{ |
|
"start": 515, |
|
"end": 542, |
|
"text": "(Talebi and Milanfar, 2017)", |
|
"ref_id": "BIBREF26" |
|
}, |
|
{ |
|
"start": 761, |
|
"end": 773, |
|
"text": "(Mane, 2017)", |
|
"ref_id": "BIBREF21" |
|
}, |
|
{ |
|
"start": 1003, |
|
"end": 1022, |
|
"text": "(Viswanathan, 2017)", |
|
"ref_id": "BIBREF28" |
|
}, |
|
{ |
|
"start": 1064, |
|
"end": 1086, |
|
"text": "(Agarwal et al., 2015)", |
|
"ref_id": "BIBREF0" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1." |
|
}, |
|
{ |
|
"text": "While the expression \"artistic genre\" is used to divide artworks according to the themes depicted (e.g. landscape, self-portrait, marine, religious, etc.), the term \"artistic style\" is used to refer to groups of works that have similar but not rigorously defined properties. This set of distinctive characteristics \"permits the grouping of artworks into related art movements\" (Bar et al., 2014) . For example, Impressionism is characterised by the use of flurried brushstrokes to represent the subject with gesture and illusion (e.g. the painter Pierre-Auguste Renoir), Expressionism uses vivid and unrealistic colors to depict the subject as it appears to the artist (e.g. Wassily Kandinsky), in Abstraction the subject is reduced to its dominant colors, shapes or patterns (e.g. Piet Mondrian) and Baroque emphasize exaggerated motion and easily interpreted detail to produce drama and exuberance (e.g. Peter Paul Rubens). Figure 1 shows 5 different art styles, along with a brief description. Art style divisions are often identified by art historians based on the experience of looking at other works of art and the historical context. However, this is not an easy task since the limits between art styles are vague or blurred. Indeed, a style can span many different painters, periods and artistic schools. For example, Goya's technique influenced both late Romanticism and Impressionism and Pablo Picasso painted in both surrealist and cubist styles.", |
|
"cite_spans": [ |
|
{ |
|
"start": 377, |
|
"end": 395, |
|
"text": "(Bar et al., 2014)", |
|
"ref_id": "BIBREF2" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 926, |
|
"end": 934, |
|
"text": "Figure 1", |
|
"ref_id": "FIGREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Traditional art styles", |
|
"sec_num": "1.1." |
|
}, |
|
{ |
|
"text": "Previous artistic styles are part of the mainstream art world, which means that they all have culture as \"an inescapable aspect of image production\" (Chadwick, 2015, p. 17) . In practical terms, this means that a painter in the mainstream is inspired by the work of those who had gone before him/her but the artist is not conscious that he/she is \"imitating\" another work of art. In contrast, there is the art created outside the boundaries of official culture or \"Anti-cultural art\" as described by Jean Dubuffet in 1949. The condition of \"non-traditional\" or, more specifically, \"outsider\" artist applies to people who have very little contact with the mainstream art world and for this reason have developed extreme unconventional ideas based on spontaneous inventions (see Figure 2a -b). We are therefore talking about psychiatric hospital patients, children, self-taught artists, people in prison or with autism, etc. The art of these \"anti-intellectual, anti-professional, anti-academic\" people (Cottom, 2003) resists analysis with traditional art criteria, while the use of non-artistic criteria such as personality features, prevents the consideration of the results of the creative process (you are looking at the person not at the work of art). This is the thinking of the Outsider art collector John Soldano, for whom \"the only way for me to honestly define outsider art is by artists\" and the arts writer Priscilla Frank who says that \"while other genres like Abstract Expressionism or Cubism denote a specific set of aesthetic guidelines or artistic traditions, the label 'outsider art' reflects more the life story and mental or emotional aptitude of the artist\" (Frank, 2017) . From a stylistic point of view, outsider artists paint obsessively repetitive images or themes (see Figure 2c ). This might indicate an attempt to overcome the \"horror vacui\" (fear of empty space), bring order to mental chaos and provide reassurance that they are in control. It could be said that the outsider's vocabulary \"oscillates back and forth between the ordered and monotonous filling of the surface of the work and the rhythmic and dynamic variation between the void and fullness of the composition\" (Raw Vision magazine). Outsider Artists paint by physical impulse rather than intellectually. For that reason, subjects such as sexuality and eroticism can erupt in the most raw, emphatic and uncontrolled way (see Figure 2d ). In some cases, the artworks appear to reveal dark desires which are not often played out in reality. These, and a number of other characteristics, make Outsider Art unattractive for a large part of the population and art historians. In a larger sense, Outsider Art label covers an expanded range of non-traditional art styles such as art brut, na\u00efve art, self-taught art, art singulier, visionary art, insane art, raw art, folk art, etc. All these form part of a continuum of artistic terms with blurred lines between them that are the tip of the iceberg of a potential task of classification of non-traditional art styles. In this article we use the terms non-traditional and outsider styles interchangeably.", |
|
"cite_spans": [ |
|
{ |
|
"start": 149, |
|
"end": 172, |
|
"text": "(Chadwick, 2015, p. 17)", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 1001, |
|
"end": 1015, |
|
"text": "(Cottom, 2003)", |
|
"ref_id": "BIBREF6" |
|
}, |
|
{ |
|
"start": 1677, |
|
"end": 1690, |
|
"text": "(Frank, 2017)", |
|
"ref_id": "BIBREF11" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 777, |
|
"end": 786, |
|
"text": "Figure 2a", |
|
"ref_id": "FIGREF1" |
|
}, |
|
{ |
|
"start": 1793, |
|
"end": 1802, |
|
"text": "Figure 2c", |
|
"ref_id": "FIGREF1" |
|
}, |
|
{ |
|
"start": 2417, |
|
"end": 2426, |
|
"text": "Figure 2d", |
|
"ref_id": "FIGREF1" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Outsider Art and non-traditional art styles", |
|
"sec_num": "1.2." |
|
}, |
|
{ |
|
"text": "Studies addressing the topic of the computational analysis of works of art are based on extracting a set of image features and using them to train different classifiers. Various formal image features such as line, color, texture or brush strokes and functional image features such as expression, content, composition and meaning (iconography) can be used to classify art styles automatically for paintings.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Classifying art style automatically in painting", |
|
"sec_num": "1.3." |
|
}, |
|
{ |
|
"text": "Classifying an artistic style automatically in painting has been the subject of much recent work that can be loosely divided into hand-crafted features and CNN-based features (training from scratch and pre-trained models). The former category (see Figure 3a) is a past tendency based in the use of computer vision methods to model handcrafted low-level features (e.g., color histograms, SIFT/GIST descriptor, texture, edges, brightness and gradient) that can be used by machine learning methods (e.g., SVM). The latter category (see Figure 3b ) is a growing tendency and uses a Convolutional Neural Network (CNN) that encodes image content (semantic features) from a very large set of data (Zhao et al., 2017) . Some examples of these two methods are briefly described below. 2.1. Handcrafted low-level features Gunsel et al. (2005) trained an SVM classifier to discriminate between five painting styles. Their system computes a 6-dimensional vector of low level features. The authors report 90% accuracy with a low number of false positives. Jiang et al. (2006) classified traditional Chinese paintings into one of the two styles, Gongbi (traditional Chinese realistic painting) or Xieyi (freehand style) by using low-level features. They reported an accuracy rate of around 90% when combining decision tree and SVMs classifiers. Wallraven et al. (2009) tested how well several low-level features describe images from 11 different art periods. The authors found than \"computational classifiers created from the participant data are able to categorize art periods with a performance of around 66%\". The overall conclusion was that images grouped by humans corresponded better with the canonical art periods than those clustered by the computer. Siddiquie et al. (2009) obtained good results in the clas-sification of seven different painting styles by using multiple kernel learning in conjunction with low-level features (with accuracy rates of 76% to 92%). Zujovic et al. (2009) reported an overall accuracy rate of 69.1% when classifying five different genres. ", |
|
"cite_spans": [ |
|
{ |
|
"start": 690, |
|
"end": 709, |
|
"text": "(Zhao et al., 2017)", |
|
"ref_id": "BIBREF31" |
|
}, |
|
{ |
|
"start": 812, |
|
"end": 832, |
|
"text": "Gunsel et al. (2005)", |
|
"ref_id": "BIBREF13" |
|
}, |
|
{ |
|
"start": 1043, |
|
"end": 1062, |
|
"text": "Jiang et al. (2006)", |
|
"ref_id": "BIBREF18" |
|
}, |
|
{ |
|
"start": 1331, |
|
"end": 1354, |
|
"text": "Wallraven et al. (2009)", |
|
"ref_id": "BIBREF29" |
|
}, |
|
{ |
|
"start": 1745, |
|
"end": 1768, |
|
"text": "Siddiquie et al. (2009)", |
|
"ref_id": "BIBREF24" |
|
}, |
|
{ |
|
"start": 1959, |
|
"end": 1980, |
|
"text": "Zujovic et al. (2009)", |
|
"ref_id": "BIBREF32" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 248, |
|
"end": 258, |
|
"text": "Figure 3a)", |
|
"ref_id": "FIGREF2" |
|
}, |
|
{ |
|
"start": 533, |
|
"end": 542, |
|
"text": "Figure 3b", |
|
"ref_id": "FIGREF2" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Related works", |
|
"sec_num": "2." |
|
}, |
|
{ |
|
"text": "In the task of classifying 25 different painting styles from the Wikipainting dataset, Karayev et al. (2014) calculated through the confusion matrix up to 0.81 accuracy at predicting the Ukiyo-e style. They also found that the DeCAF, a deep CNN originally trained for object recognition, performs best for the task of classifying novel images according to their style. This leads them to conclude that some styles are closely related to image content, that is, the existence of certain objects in the painting. Bar et al. (2014) examine binarized features derived from a Deep Neural Network in order to identify the style of paintings. They apply PiCoDes (\"Picture Codes\"), a very compact image descriptor, to learn a compact binary representation of an image. Their baseline was extracted from a CNN trained on the ImageNet dataset and implemented in Decaf, a deep convolutional activation feature for generic visual recognition. Their results show an improvement in performance with CNN-based features (0.43% accuracy) as well as their binarized version to distinguish 27 painting styles compared to hand-crafted low level descriptors (0.37% accuracy) such as Edge texture information and color histogram. Mao et al. (2017) implemented DeepArt, a unified framework that can learn simultaneously both the contents and style of visual arts from a large number of digital artworks with multi-labels. The architecture of the framework is constructed by dual feature extraction paths that can extract style features and content features, respectively. The content feature representation path is generated on the basis of a VGG-16 network and the style feature representation path is built by adopting a Gram matrix to the filter responses in certain layers of the VGG-16 network. According to the authors, embedding the two output features in a single representation can be used to further improve two tasks: the automatic retrieval and annotation of digital artworks. With the goal of outperforming the state-of-the-art, Hong and Kim (2017) trained a CNN on an art painting dataset of 30,000 distorted (projected, rotated, scaled, etc.) images to simulate real-world displaying conditions. Three different architectures of CNN were tested on this dataset: the first architecture was derived from AlexNet (Krizhevsky et al., 2012) , the second architecture was inspired by VGGNet (Simonyan and Zisserman, 2014) and the third architecture was a smaller version of the second one which used a smaller filter size (11 \u2192 7) in the beginning and fewer neurons in fully-connected layers. The latter architecture performed best, obtaining low test error rates by optimizing its parameters with the Adam algorithm. According to the researchers, the proposed CNN-based method outperformed the previous state-of-the-art with a test error rate of 15.6% to 2%.", |
|
"cite_spans": [ |
|
{ |
|
"start": 87, |
|
"end": 108, |
|
"text": "Karayev et al. (2014)", |
|
"ref_id": "BIBREF19" |
|
}, |
|
{ |
|
"start": 511, |
|
"end": 528, |
|
"text": "Bar et al. (2014)", |
|
"ref_id": "BIBREF2" |
|
}, |
|
{ |
|
"start": 1208, |
|
"end": 1225, |
|
"text": "Mao et al. (2017)", |
|
"ref_id": "BIBREF22" |
|
}, |
|
{ |
|
"start": 2302, |
|
"end": 2327, |
|
"text": "(Krizhevsky et al., 2012)", |
|
"ref_id": "BIBREF20" |
|
}, |
|
{ |
|
"start": 2377, |
|
"end": 2407, |
|
"text": "(Simonyan and Zisserman, 2014)", |
|
"ref_id": "BIBREF25" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "CNN-based features", |
|
"sec_num": "2.2." |
|
}, |
|
{ |
|
"text": "In order to identify the best training setup for the style classification of paintings, Cetinic et al. (2018) compared different CNN fine-tuning strategies performed on a WikiArt subset of 27 classes in which each class contains more than 800 paintings. They used visual features (e.g. edges or blobs) and content features (e.g. scenes and objects in paintings) derived from the layers of a CNN pre-trained on the ImageNet dataset (CaffeNet). Overall results indicate a lower accuracy for style classification due to the overlapping of visual properties between classes and the great diversity of content depicted in each style. The most distinctively categorized style was Ukiyo-e (84%) and the least distinctive was Academism, which was misclassified. On the basis of these results, researchers conclude that style is not only associated with mere visual characteristics and the content of paintings, but is often a contextually dependent concept. ", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "CNN-based features", |
|
"sec_num": "2.2." |
|
}, |
|
{ |
|
"text": "Previous research has reported heterogeneous performances for the style classification of fine art paintings, depending on the type of features used and the number of categories created. Nevertheless, there is a significant degree of agreement on the prevalence of binarized features derived from a deep neural network over hand-crafted low level descriptors. But, can these findings be considered valid for non-traditional artistic styles? Such a question arises due to the fact that, as described in the Introduction, non-traditional styles are influenced by factors \"outside\" of the world of art. Additionally, Florea et al. (2016) showed that several artistic styles resist certain classification techniques.", |
|
"cite_spans": [ |
|
{ |
|
"start": 614, |
|
"end": 634, |
|
"text": "Florea et al. (2016)", |
|
"ref_id": "BIBREF10" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Preliminary experiments", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "Two different experiments were conducted in order to achieve a first approach to the classification of nontraditional styles. These experiments perform the binary and multiclass classification of Outsider Art and traditional styles.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Preliminary experiments", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "To study the performance correlation between Outsider Art and traditional styles, we trained, validated, and tested different networks using images from WikiArt and Outsider Art datasets. WikiArt is the largest public available dataset and contains 82,653 images classified in 27 artistic styles. It is fair to note that: (i) the \"Wikiart collection [...] contains various paintings from different styles that are erroneously labeled\" (Elgammal et al., 2018, p. 6 ) and (ii) this is an unbalanced dataset as seen in Figure 4 . For its part, the Outsider Art dataset merges 2,405 images labeled as Na\u00efve Art from WikiArt, which is considered very close to the Outsider Art style (Van Heddeghem, 2016, p. 13), and 1,232 Outsider Art images collected specifically for this paper (in total 3,616 images). In the experiments, the number of images and classes was reduced in order to work with balanced data. ", |
|
"cite_spans": [ |
|
{ |
|
"start": 435, |
|
"end": 463, |
|
"text": "(Elgammal et al., 2018, p. 6", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 516, |
|
"end": 524, |
|
"text": "Figure 4", |
|
"ref_id": "FIGREF3" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Experimental setup", |
|
"sec_num": "3.1." |
|
}, |
|
{ |
|
"text": "This experiment aims at answering the following scientific question: Does the Outsider style show a performance in the task of classifying paintings comparable to those of the In this regard, WikiArt and Outsider Art datasets were used as basis categories for mapping the problem to multiple binary classification tasks (e.g. Cubism versus Outsider Art). Datasets were balanced by selecting 2,561 images per class and merging similar styles in ten basic categories: Cubism (CUB), Baroque (BAR), Abstract (ABS), Renaissance (REN), Romanticism (ROM), Expressionism (EXP), Modern Art (MOD), Realism (REA), Impressionism (IMP) and Outsider Art (OUT). As a result, the final dataset included 10 categories and 25,610 images that were resized to 28 \u00d7 21 pixels. We trained several Convolutional Neural Networks from scratch using Keras API with Tensorflow as backend. Accuracies were obtained with 100 epochs because our tests indicate that using a number of epochs greater than 100 does not increase the performance significantly. Additionally, as is usually done in the literature, 70% of data were used for training, 20% of data for validation and 10% for testing. During the classification, all styles were crossed with each other in order to obtain the accuracies listed in Figure 5 . The left hand column in the same Figure contains the average accuracy (%) obtained by each style. In general, our results suggest that the task of classifying the Outsider Art style does not differ from classifying traditional styles. The classification of Outsider Art achieves a general average accuracy of 72,04%, which is below the average for Baroque (77,69%) and above the average for Expressionism (68,10%). This may indicate that, in contrast to what art historians state, this so-called anti-cultural art can be analyzed under the same parameters and conditions as mainstream art. These preliminary results also show that Outsider Art is closely related to Cubism, Expressionism and Modern Art, resulting in poor accuracies (62.3%, 62.5% and 67.5%, respectively). Indeed, these three styles of art present the lowest average accuracy levels of the entire classification (70.9%, 69.5% and 68.1%, respectively). These analyses further show that while it is relatively easy for the classifier to differentiate Outsider Art from Baroque (82.2% of accuracy), Cubism and Expressionism are the pair of traditional styles that are more difficult to classify (58% of accuracy),", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 1273, |
|
"end": 1281, |
|
"text": "Figure 5", |
|
"ref_id": "FIGREF4" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Classification from scratch", |
|
"sec_num": "3.2." |
|
}, |
|
{ |
|
"text": "while Baroque is the easiest style to classify. However, although this seems obvious, it is important to emphasise that while style classification accuracies between pairs of styles are high (the estimated average efficiency levels are about 72,49%, see Figure 5 ), test accuracies drop dramatically when we classify three or more categories under a basic configuration: 3 styles/categories (62,4%), 4 styles/categories (52,2%), and so on, until the 10 styles (23,6%). In other words, it is essential to find features which can discriminate among multiple artistic styles. The second part of the following experiment tackles this issue.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 254, |
|
"end": 262, |
|
"text": "Figure 5", |
|
"ref_id": "FIGREF4" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Classification from scratch", |
|
"sec_num": "3.2." |
|
}, |
|
{ |
|
"text": "This second experiment aims at answering the following scientific question: Is it possible to improve accuracy for Outsider Art style classification by using pre-trained models? Pre-trained models, such as ImageNet, VGGNet and ResNet use fine-tuned features that were originally trained on a different but correlated problem, to match the current problem. We trained a ResNet-18 model 1 to perform a binary classification problem: traditional versus Outsider Art styles. The dataset used is balanced, containing 2,028 images, 1,014 for Outsider Art and 1,014 for traditional art (homogeneously sampling 39 images for each of the 26 styles of traditional art from WikiArt). The test set is also balanced and has 416 images (208 for each category with the same sampling of 8 images for each of the 26 styles). The training is done for 40 epochs using pretrained weights from ImageNet 2 . The batch size used is 128, weight decay 5e-4, momentum 0.9. Learning rate for blocks 1, 2 and 3 of ResNet-18 is set to 0.001 and the block 4 and the classifiers have a learning rate of 0.01. All learning rates are multiplied by 0.1 at epochs 20, 30 and 35. The model selected for classification is the one in the last epoch as no validation set was built due to the lack of data. The loss used is binary cross-entropy. Under this fine-tuned configuration, an accuracy of 84.3% was achieved. This accuracy outperforms all previous accuracies based on CNN trained from scratch, which means that repurposing and fine-tuning features can be used to obtain better feature rep-resentations of the Outsider Art style.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Classification using a pre-trained model", |
|
"sec_num": "3.3." |
|
}, |
|
{ |
|
"text": "We have also trained a ResNet-18 model to perform a multiclass classification problem. The loss used is crossentropy. The dataset is the same as that used in the experiment described in Section 3.2. with 224x224 crops extracted from the images resized to 256 in the smallest side (preserving the aspect ratio). The training was done for 120 epochs using pre-trained weights from ImageNet. Batch size used is 256, weight decay 5e-4, momentum 0.9. Initial learning rates are: for classifier 0.01, for blocks 3 and 4 0.001 and the rest of the parameters 0.0001. All learning rates are multiplied by 0.1 at epochs 70 and 100. The model selected for classification is the one in the best epoch of validation, whose accuracy is 61.17188 (see Figure 6) . The test accuracy is 61.9141 (per class, Abstract: 0.9414, Baroque: 0.7891, Cubism: 0.8672, Expressionism: 0.5508, Impressionism: 0.5938, Modern: 0.5938, Outsider: 0.7695, Realism: 0.5273, Renaissance: 0.7695, Romanticism: 0.8047). This result agrees with the results from researchers in section 2.2., showing once again that there are no significant differences in classifying traditional and non-traditional art styles. ", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 736, |
|
"end": 745, |
|
"text": "Figure 6)", |
|
"ref_id": "FIGREF5" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Classification using a pre-trained model", |
|
"sec_num": "3.3." |
|
}, |
|
{ |
|
"text": "This position paper has analysed the possibility of classifying non-traditional artistic styles by using the same binarized features that are used for classifying traditional styles. The first part of the paper introduces the theoretical elements that constitute a framework for understanding the problem. The second part describes the state-of-the-art on classifying art styles automatically in paintings. Due to the good accuracy performance of Deep Learning-based methods for classifying traditional art styles, it was suggested to apply them to classify non-traditional art styles (i.e. Outsider Art). Our preliminary experiments have provided good reasons to think that, as is the case with traditional styles, the Outsider Art can be computationally modelled by objective means. Additionally, in accordance with theoretical (Frank, 2017) and applied (Yang et al., 2018) studies, we assume that the automatic classification of the Outsider Art style should consider a multimodal approach based on an analysis of images, as well as text. From our point of view, this twofold strategy will involve (i) the compilation of a big dataset of Outsider Art images and (ii) the creation of an ontology of Outsider Art. On the one hand, the image dataset will contain thousands of digital paintings in the Outsider Art style that can be used by machine learning algorithm. This resource can be integrated with the Outsider Art ontology to obtain a multimodal dataset for understanding Outsider Art, similar to that suggested by (Garcia and Vogiatzis, 2019) . On the other hand, the Outsider Art ontology will focus on representing part of our existing knowledge of this artistic style in a machine-readable language. A particular feature of the Outsider Art knowledge is that it includes both aesthetic entities and social/medical issues, for example: \"(Gaston Chaissac) suffered from tuberculosis, and for a time, produced art while convalescing in a sanitorium\" (Wikipedia). Therefore, the source text that we will use for ontology learning is a representative set of scientific books, papers, magazines and web pages. Additionally, we will integrate in our model some existing ontologies and terminologies such as the Conceptual Reference Model (CIDOC CRM) (Le Boeuf et al., 2019), the Europeana Data Model (EDM) (Europeana, 2017) , the Art & Architecture Thesaurus (Alexiev et al., 2017) ), the Cultural Objects Name Authority (CONA) (Harpring, 2019a) , the Getty Iconography Authority (AI) (Harpring, 2019b) and the Getty Union List of Artist Names (ULAN) (Harpring, 2019c) . Currently, we are in the first phase of the project and aim to semi-automatically construct an exhaustive corpus that consists of semantically tagged texts. Our purpose is to apply this corpus to the construction of a large-scale corpus through the automatic retrieval and annotation of new texts. In the second phase of the project, we will extract the ontology from the corpus and we will use the ontology for automatic image annotation and retrieval.", |
|
"cite_spans": [ |
|
{ |
|
"start": 830, |
|
"end": 843, |
|
"text": "(Frank, 2017)", |
|
"ref_id": "BIBREF11" |
|
}, |
|
{ |
|
"start": 1523, |
|
"end": 1551, |
|
"text": "(Garcia and Vogiatzis, 2019)", |
|
"ref_id": "BIBREF12" |
|
}, |
|
{ |
|
"start": 2311, |
|
"end": 2328, |
|
"text": "(Europeana, 2017)", |
|
"ref_id": "BIBREF9" |
|
}, |
|
{ |
|
"start": 2364, |
|
"end": 2386, |
|
"text": "(Alexiev et al., 2017)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 2433, |
|
"end": 2450, |
|
"text": "(Harpring, 2019a)", |
|
"ref_id": "BIBREF14" |
|
}, |
|
{ |
|
"start": 2490, |
|
"end": 2507, |
|
"text": "(Harpring, 2019b)", |
|
"ref_id": "BIBREF15" |
|
}, |
|
{ |
|
"start": 2556, |
|
"end": 2573, |
|
"text": "(Harpring, 2019c)", |
|
"ref_id": "BIBREF16" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion and future work", |
|
"sec_num": "4." |
|
}, |
|
{ |
|
"text": "https://arxiv.org/abs/1512.03385 2 https://ieeexplore.ieee.org/document/5206848", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [ |
|
{ |
|
"text": "This work has received funding from the Irish Research Council (Grant GOIPD/2019/463) and the ADAPT Centre for Digital Content Technology which is funded under the SFI Research Centres Programme (Grant 13/RC/2106). This research has been also supported by Science Foundation Ireland under grant number SFI/15/SIRG/3283.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Acknowledgements", |
|
"sec_num": "5." |
|
} |
|
], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "Genre and style based painting classification", |
|
"authors": [ |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Agarwal", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Karnick", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "N", |
|
"middle": [], |
|
"last": "Pant", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "U", |
|
"middle": [], |
|
"last": "Patel", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "2015 IEEE Winter Conference on Applications of Computer Vision", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "588--594", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Agarwal, S., Karnick, H., Pant, N., and Patel, U. (2015). Genre and style based painting classification. In 2015 IEEE Winter Conference on Applications of Computer Vision, pages 588-594, Waikoloa Beach, Hawaii, Jan- uary. IEEE.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "Getty Vocabularies: Linked Open Data version 3.4. Semantic Representation. The Getty Vocabularies", |
|
"authors": [ |
|
{ |
|
"first": "V", |
|
"middle": [], |
|
"last": "Alexiev", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Cobb", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "G", |
|
"middle": [], |
|
"last": "Garcia", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Harpring", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Alexiev, V., Cobb, J., Garcia, G., and Harpring, P., (2017). Getty Vocabularies: Linked Open Data version 3.4. Se- mantic Representation. The Getty Vocabularies.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Classification of artistic styles using binarized features derived from a deep neural network", |
|
"authors": [ |
|
{ |
|
"first": "Y", |
|
"middle": [], |
|
"last": "Bar", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "N", |
|
"middle": [], |
|
"last": "Levy", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "Wolf", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Bar, Y., Levy, N., and Wolf, L. (2014). Classification of artistic styles using binarized features derived from a deep neural network. Computer Vision (Lecture Notes in Computer Science), 8925.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Fine-tuning convolutional neural networks for fine art classification", |
|
"authors": [ |
|
{ |
|
"first": "E", |
|
"middle": [], |
|
"last": "Cetinic", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "Lipic", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Grgic", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "Expert Systems with Applications", |
|
"volume": "114", |
|
"issue": "", |
|
"pages": "107--118", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Cetinic, E., Lipic, T., and Grgic, S. (2018). Fine-tuning convolutional neural networks for fine art classification. Expert Systems with Applications, 114:107-118.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Disorienting Forms: Jean Dubuffet", |
|
"authors": [ |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Chadwick", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Chadwick, S. (2015). Disorienting Forms: Jean Dubuf- fet, Portraiture, Ethnography. Rice University, Houston, Texas.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "Automatically classifying paintings with perceptual inspired descriptors", |
|
"authors": [ |
|
{ |
|
"first": "R", |
|
"middle": [ |
|
"G" |
|
], |
|
"last": "Condorovici", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Florea", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Vertan", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "Journal of Visual Communication and Image Representation", |
|
"volume": "26", |
|
"issue": "", |
|
"pages": "222--230", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Condorovici, R. G., Florea, C., and Vertan, C. (2015). Automatically classifying paintings with perceptual in- spired descriptors. Journal of Visual Communication and Image Representation, 26:222-230.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "Why Education Is Useless", |
|
"authors": [ |
|
{ |
|
"first": "D", |
|
"middle": [], |
|
"last": "Cottom", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Cottom, D. (2003). Why Education Is Useless. Phd. thesis, University of Pennsylvania Press.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "Classification of art paintings by genre", |
|
"authors": [ |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Culjak", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "B", |
|
"middle": [], |
|
"last": "Miku\u0161", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "K", |
|
"middle": [], |
|
"last": "Je\u017e", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Hadjic", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2011, |
|
"venue": "Proceedings of the 34th International Convention", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "345--369", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Culjak, M., Miku\u0161, B., Je\u017e, K., and Hadjic, S. (2011). Classification of art paintings by genre. In Proceedings of the 34th International Convention, pages 345--369, Opatija: IEEE.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "The shape of art history in the eyes of the machine", |
|
"authors": [ |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Elgammal", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Mazzone", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "B", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "D", |
|
"middle": [], |
|
"last": "Kim", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Elhoseiny", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "32nd AAAI conference on Artificial Intelligence", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Elgammal, A., Mazzone, M., Liu, B., Kim, D., and Elho- seiny, M. (2018). The shape of art history in the eyes of the machine. In 32nd AAAI conference on Artificial Intelligence, New Orleans, USA.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "Definition of the Europeana Data Model v5.2.8", |
|
"authors": [ |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Europeana", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Europeana, (2017). Definition of the Europeana Data Model v5.2.8. European Union.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "Pandora: Description of a painting database for art movement recognition with baselines and perspectives", |
|
"authors": [ |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Florea", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Condorovici", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Vertan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Butnaru", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "Florea", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Vr\u01ceanceanu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "Proceedings of the European Signal Processing Conference (EUSIPCO)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Florea, C., Condorovici, R., Vertan, C., Butnaru, R., Florea, L., and Vr\u01ceanceanu, R. (2016). Pandora: Description of a painting database for art movement recognition with baselines and perspectives. Proceedings of the European Signal Processing Conference (EUSIPCO).", |
|
"links": null |
|
}, |
|
"BIBREF11": { |
|
"ref_id": "b11", |
|
"title": "What is the meaning of outsider art? the genre with a story", |
|
"authors": [ |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Frank", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Frank, P. (2017). What is the meaning of outsider art? the genre with a story, not a style.", |
|
"links": null |
|
}, |
|
"BIBREF12": { |
|
"ref_id": "b12", |
|
"title": "How to read paintings: Semantic art understanding with multi-modal retrieval", |
|
"authors": [ |
|
{ |
|
"first": "N", |
|
"middle": [], |
|
"last": "Garcia", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "G", |
|
"middle": [], |
|
"last": "Vogiatzis", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "Computer Vision -ECCV 2018 Workshops, Proceedings, volume 11130 of Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "676--691", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Garcia, N. and Vogiatzis, G. (2019). How to read paint- ings: Semantic art understanding with multi-modal re- trieval. In Stefan Roth et al., editors, Computer Vision -ECCV 2018 Workshops, Proceedings, volume 11130 of Lecture Notes in Computer Science (including sub- series Lecture Notes in Artificial Intelligence and Lec- ture Notes in Bioinformatics), pages 676-691, Germany, 1. Springer.", |
|
"links": null |
|
}, |
|
"BIBREF13": { |
|
"ref_id": "b13", |
|
"title": "Contentbased access to art paintings", |
|
"authors": [ |
|
{ |
|
"first": "B", |
|
"middle": [], |
|
"last": "Gunsel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Sariel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "O", |
|
"middle": [], |
|
"last": "Icoglu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "", |
|
"volume": "2", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Gunsel, B., Sariel, S., and Icoglu, O. (2005). Content- based access to art paintings. volume 2, pages II -558, 10.", |
|
"links": null |
|
}, |
|
"BIBREF14": { |
|
"ref_id": "b14", |
|
"title": "Cultural Objects Name Authority (CONA): Introduction and Overview", |
|
"authors": [ |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Harpring", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Harpring, P., (2019a). Cultural Objects Name Authority (CONA): Introduction and Overview. Getty Vocabulary Program.", |
|
"links": null |
|
}, |
|
"BIBREF15": { |
|
"ref_id": "b15", |
|
"title": "The Getty Iconography Authority: Introduction and Overview", |
|
"authors": [ |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Harpring", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Harpring, P., (2019b). The Getty Iconography Authority: Introduction and Overview. Getty Vocabulary Program.", |
|
"links": null |
|
}, |
|
"BIBREF16": { |
|
"ref_id": "b16", |
|
"title": "The Getty Union List of Artist Names: Introduction and Overview", |
|
"authors": [ |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Harpring", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Harpring, P., (2019c). The Getty Union List of Artist Names: Introduction and Overview. Getty Vocabulary Program.", |
|
"links": null |
|
}, |
|
"BIBREF17": { |
|
"ref_id": "b17", |
|
"title": "Art painting identification using convolutional neural network", |
|
"authors": [ |
|
{ |
|
"first": "Y", |
|
"middle": [], |
|
"last": "Hong", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Kim", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "International Journal of Applied Engineering Research", |
|
"volume": "12", |
|
"issue": "", |
|
"pages": "532--539", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Hong, Y. and Kim, J. (2017). Art painting identification using convolutional neural network. International Jour- nal of Applied Engineering Research, 12:532-539.", |
|
"links": null |
|
}, |
|
"BIBREF18": { |
|
"ref_id": "b18", |
|
"title": "An effective method to detect and categorize digitized traditional chinese paintings", |
|
"authors": [ |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Jiang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Q", |
|
"middle": [], |
|
"last": "Huang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Q", |
|
"middle": [], |
|
"last": "Ye", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "W", |
|
"middle": [], |
|
"last": "Gao", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "Pattern Recognition Letters", |
|
"volume": "27", |
|
"issue": "", |
|
"pages": "734--746", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jiang, S., Huang, Q., Ye, Q., and Gao, W. (2006). An effective method to detect and categorize digitized tra- ditional chinese paintings. Pattern Recognition Letters, 27:734--746.", |
|
"links": null |
|
}, |
|
"BIBREF19": { |
|
"ref_id": "b19", |
|
"title": "Recognizing image style", |
|
"authors": [ |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Karayev", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Hertzmann", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Winnemoeller", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Agarwala", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Darrell", |
|
"middle": [], |
|
"last": "", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the British Machine Vision Conference", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Karayev, S., Hertzmann, A., Winnemoeller, H., Agarwala, A., and Darrell, T. (2014). Recognizing image style. In Proceedings of the British Machine Vision Conference, Nottingham, England. BMVA Press.", |
|
"links": null |
|
}, |
|
"BIBREF20": { |
|
"ref_id": "b20", |
|
"title": "Imagenet classification with deep convolutional neural networks", |
|
"authors": [ |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Krizhevsky", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "I", |
|
"middle": [], |
|
"last": "Sutskever", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "G", |
|
"middle": [], |
|
"last": "Hinton", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2012, |
|
"venue": "Advances in Neural Information Processing Systems", |
|
"volume": "25", |
|
"issue": "", |
|
"pages": "1106--1114", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Krizhevsky, A., Sutskever, I., and Hinton, G. (2012). Im- agenet classification with deep convolutional neural net- works. Advances in Neural Information Processing Sys- tems, 25:1106-1114.", |
|
"links": null |
|
}, |
|
"BIBREF21": { |
|
"ref_id": "b21", |
|
"title": "Detection of forgery in art paintings using machine learning", |
|
"authors": [ |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Mane", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "International Journal of Innovative Research in Science, Engineering and Technology", |
|
"volume": "6", |
|
"issue": "", |
|
"pages": "8681--8692", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Mane, S. (2017). Detection of forgery in art paintings us- ing machine learning. International Journal of Innova- tive Research in Science, Engineering and Technology, 6:8681-8692.", |
|
"links": null |
|
}, |
|
"BIBREF22": { |
|
"ref_id": "b22", |
|
"title": "Deepart: Learning joint representations of visual arts", |
|
"authors": [ |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Mao", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Cheung", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "She", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "Proceedings of the 25th ACM international conference on Multimedia", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Mao, H., Cheung, M., and She, J. (2017). Deepart: Learn- ing joint representations of visual arts. In Proceedings of the 25th ACM international conference on Multimedia, Mountain View California, USA.", |
|
"links": null |
|
}, |
|
"BIBREF23": { |
|
"ref_id": "b23", |
|
"title": "Impressionism, expressionism, surrealism: automated recognition of painters and schools of art", |
|
"authors": [ |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "Shamir", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "Macura", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "N", |
|
"middle": [], |
|
"last": "Orlov", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "D", |
|
"middle": [ |
|
"M" |
|
], |
|
"last": "Eckley", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "I", |
|
"middle": [ |
|
"G" |
|
], |
|
"last": "Goldberg", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "ACM Transactions on Applied Perception (TAP)", |
|
"volume": "7", |
|
"issue": "", |
|
"pages": "1--18", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Shamir, L., Macura, T., Orlov, N., Eckley, D. M., and Gold- berg, I. G. (2010). Impressionism, expressionism, sur- realism: automated recognition of painters and schools of art. ACM Transactions on Applied Perception (TAP), 7:1--18.", |
|
"links": null |
|
}, |
|
"BIBREF24": { |
|
"ref_id": "b24", |
|
"title": "Combining multiple kernels for efficient image classification", |
|
"authors": [ |
|
{ |
|
"first": "B", |
|
"middle": [], |
|
"last": "Siddiquie", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [ |
|
"N" |
|
], |
|
"last": "Vitaladevuni", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "L", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Davis", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Workshop on the Applications of Computer Vision (WACV)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Siddiquie, B., Vitaladevuni, S. N., and Davis, L. S. (2009). Combining multiple kernels for efficient image classifi- cation. In Workshop on the Applications of Computer Vi- sion (WACV), Snowbird, UT.", |
|
"links": null |
|
}, |
|
"BIBREF25": { |
|
"ref_id": "b25", |
|
"title": "Very deep convolutional networks for large-scale image recognition", |
|
"authors": [ |
|
{ |
|
"first": "K", |
|
"middle": [], |
|
"last": "Simonyan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Zisserman", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "3rd International Conference on Learning Representations", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Simonyan, K. and Zisserman, A. (2014). Very deep convo- lutional networks for large-scale image recognition. In 3rd International Conference on Learning Representa- tions (ICLR 2015), San Diego, CA.", |
|
"links": null |
|
}, |
|
"BIBREF26": { |
|
"ref_id": "b26", |
|
"title": "NIMA: Neural Image Assessment. Institute of Electrical and Electronics Engineers", |
|
"authors": [ |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Talebi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Milanfar", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Talebi, H. and Milanfar, P. (2017). NIMA: Neural Image Assessment. Institute of Electrical and Electronics Engi- neers (IEEE), San Diego, CA.", |
|
"links": null |
|
}, |
|
"BIBREF27": { |
|
"ref_id": "b27", |
|
"title": "Outsider Art, In or Outside the World of Art? A study of the framing of the paradoxical position of outsider art", |
|
"authors": [ |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Van Heddeghem", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Van Heddeghem, R. (2016). Outsider Art, In or Outside the World of Art? A study of the framing of the para- doxical position of outsider art. Master thesis, Erasmus School of History, Culture and Communication, Erasmus University Rotterdam.", |
|
"links": null |
|
}, |
|
"BIBREF28": { |
|
"ref_id": "b28", |
|
"title": "Artist identification with convolutional neural networks", |
|
"authors": [ |
|
{ |
|
"first": "N", |
|
"middle": [], |
|
"last": "Viswanathan", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Viswanathan, N. (2017). Artist identification with con- volutional neural networks. Technical report, Stanford University, California, USA.", |
|
"links": null |
|
}, |
|
"BIBREF29": { |
|
"ref_id": "b29", |
|
"title": "Categorizing art: comparing humans and computers", |
|
"authors": [ |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Wallraven", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Fleming", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "D", |
|
"middle": [], |
|
"last": "Cunningham", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Rigau", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "F", |
|
"middle": [ |
|
"M" |
|
], |
|
"last": "Sbert", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Computers and Graphics", |
|
"volume": "33", |
|
"issue": "", |
|
"pages": "484--495", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Wallraven, C., Fleming, R., Cunningham, D., Rigau, J., F. M., and Sbert, M. (2009). Categorizing art: compar- ing humans and computers. Computers and Graphics, 33:484-495.", |
|
"links": null |
|
}, |
|
"BIBREF30": { |
|
"ref_id": "b30", |
|
"title": "Historical context-based style classification of painting images via label distribution learning", |
|
"authors": [ |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Yang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "Chen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "Zhang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "X", |
|
"middle": [], |
|
"last": "Sun", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "D", |
|
"middle": [], |
|
"last": "She", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Lu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Cheng", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "Proceedings of the 26th ACM international conference on Multimedia (MM '18)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1154--1162", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Yang, J., Chen, L., Zhang, L., Sun, X., She, D., Lu, S., and Cheng, M. (2018). Historical context-based style classi- fication of painting images via label distribution learning. In Proceedings of the 26th ACM international conference on Multimedia (MM '18), pages 1154-1162, New York, NY, USA.", |
|
"links": null |
|
}, |
|
"BIBREF31": { |
|
"ref_id": "b31", |
|
"title": "Learning semantic feature map for visual content recognition", |
|
"authors": [ |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Zhao", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Z", |
|
"middle": [], |
|
"last": "Wu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Y", |
|
"middle": [], |
|
"last": "Jiang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "Proceedings of the 25th ACM international conference on Multimedia", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1291--1299", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Zhao, R., Wu, Z., Li, J., and Jiang, Y. (2017). Learning semantic feature map for visual content recognition. In Proceedings of the 25th ACM international conference on Multimedia, pages 1291-1299, New York, NY, USA. Association for Computing Machinery.", |
|
"links": null |
|
}, |
|
"BIBREF32": { |
|
"ref_id": "b32", |
|
"title": "Classifying paintings by artistic genre: an analysis of features and classifiers", |
|
"authors": [ |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Zujovic", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "Gandy", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Friedman", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "B", |
|
"middle": [], |
|
"last": "Pardo", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "T", |
|
"middle": [ |
|
"N" |
|
], |
|
"last": "Pappas", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "International Workshop on Multimedia Signal Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Zujovic, J., Gandy, L., Friedman, S., Pardo, B., and Pappas, T. N. (2009). Classifying paintings by artistic genre: an analysis of features and classifiers. In International Workshop on Multimedia Signal Processing, Rio De Janeiro: IEEE.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"FIGREF0": { |
|
"uris": null, |
|
"text": "Some examples of traditional artistic styles.", |
|
"type_str": "figure", |
|
"num": null |
|
}, |
|
"FIGREF1": { |
|
"uris": null, |
|
"text": "Some examples of Outsider Art paintings.", |
|
"type_str": "figure", |
|
"num": null |
|
}, |
|
"FIGREF2": { |
|
"uris": null, |
|
"text": "(a) low-level features (adapted from Condorovici et al. (2015) and (b) semantic features: object detection.", |
|
"type_str": "figure", |
|
"num": null |
|
}, |
|
"FIGREF3": { |
|
"uris": null, |
|
"text": "Original distribution of 27 styles: 26 traditional styles from WikiArt and the Outsider Art style (in the upper-right corner).", |
|
"type_str": "figure", |
|
"num": null |
|
}, |
|
"FIGREF4": { |
|
"uris": null, |
|
"text": "Accuracies between pairs of classes/styles. traditional styles? To answer this question, we trained several Convolutional Neural Networks to classify different pairs of traditional and non-traditional artistic styles.", |
|
"type_str": "figure", |
|
"num": null |
|
}, |
|
"FIGREF5": { |
|
"uris": null, |
|
"text": "Validation and training accuracies with respect to the Epochs.", |
|
"type_str": "figure", |
|
"num": null |
|
}, |
|
"TABREF1": { |
|
"num": null, |
|
"text": "They used the AdaBoost classifier and, as features, steerable filters, as well as edge information extracted by a canny edge detector.Shamir et al. (2010) achieved an accuracy of 91.0% by using a set of lowlevel features on paintings by nine artists working in three different styles.Culjak et al. (2011) reported a 60.2% accuracy rate in the classification of six styles (including Na\u00efve Art). They chose texture and color as low-level features and tested a range of classifiers, such as SVM.", |
|
"content": "<table><tr><td>Condorovici et al. (2015) achieved an overall detection</td></tr><tr><td>rate of 72.24% on a database containing 4119 images from</td></tr><tr><td>8 painting styles (SVM). The authors selected features rele-</td></tr><tr><td>vant for human perception and assessed the contribution of</td></tr><tr><td>each feature. The overall conclusion is that the Dominant</td></tr><tr><td>Color Volume features play a more important role for the</td></tr><tr><td>automatic identification of artistic style.</td></tr></table>", |
|
"html": null, |
|
"type_str": "table" |
|
} |
|
} |
|
} |
|
} |