• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Relative attribute based incremental learning for image recognition

    2017-05-16 10:26:43EmrahErgul

    Emrah Ergul

    Electronics&Communication Eng.Dept.of Kocaeli University,Umuttepe Campus,41380 Kocaeli,Turkey

    Original article

    Relative attribute based incremental learning for image recognition

    Emrah Ergul

    Electronics&Communication Eng.Dept.of Kocaeli University,Umuttepe Campus,41380 Kocaeli,Turkey

    A R T I C L E I N F O

    Article history:

    Received 6 November 2016

    Received in revised form

    15 January 2017

    Accepted 19 January 2017

    Available online 30 January 2017

    Image classi fication

    In this study,we propose an incremental learning approach based on a machine-machine interaction via relative attribute feedbacks that exploit comparative relationships among top level image categories.One machine acts as‘Student(S)’with initially limited information and it endeavors to capture the task domain gradually by questioning its mentor on a pool of unlabeled data.The other machine is‘Teacher (T)’with the implicit knowledge for helping S on learning the class models.T initiates relative attributes as a communication channel by randomly sorting the classes on attribute space in an unsupervised manner.S starts modeling the categories in this intermediate level by using only a limited number of labeled data.Thereafter,it first selects an entropy-based sample from the pool of unlabeled data and triggers the conversation by propagating the selected image with its belief class in a query.Since T already knows the ground truth labels,it not only decides whether the belief is true or false,but it also provides an attribute-based feedback to S in each case without revealing the true label of the query sample if the belief is false.So the number of training data is increased virtually by dropping the falsely predicted sample back into the unlabeled pool.Next,S updates the attribute space which,in fact,has an impact on T's future responses,and then the category models are updated concurrently for the next run. We experience the weakly supervised algorithm on the real world datasets of faces and natural scenes in comparison with direct attribute prediction and semi-supervised learning approaches,and a noteworthy performance increase is achieved.

    ?2017 Chongqing University of Technology.Production and hosting by Elsevier B.V.This is an open access article under the CC BY-NC-ND license(http://creativecommons.org/licenses/by-nc-nd/4.0/).

    1.Introduction

    Instead of mapping low level features(color/edge histograms, bag of words,Fourier,Wavelet transform features etc.)directly to the class labels,attribute-based approaches are nowadays popular instruments in the computer vision applications[1-8].Attributes are mainly human interpretable and visually detectable entities on the visual data.They are used for intermediate level data representations between machine-alone understandable low level features and the human speci fic top level class labels.Attributes are introduced mostly because of sparse category labeled training data and large variations among low level features caused by both viewing conditions and high intra-class variance[1,9].

    Since many top level categories share attributes as the common platforms,they can be learned with much more data samples than class labels.In addition to traditional problems like classi fication and retrieval[2,6],this leads to transferring the previously learned visual attributes to learn/detect a novel class which has not been seen during the training stage but we still know its attribute values, called zero-shot learning[10,11].Besides,we can even infer an unfamiliar class which we have not known its attribute groundtruths[12,13]at all.We may also produce more semantic queries by using the visual attributes to perceive scenes/objects in the image search engines[14].

    In this work,we propose a new approach which tries to recover from aforementioned insuf ficiencies of the supervision caused by the human factor,and the lack of connections in the attribute-based class modeling.It is based on the relative attributes for incremental visual recognition.We generate the relative attributes in an unsupervised manner by randomly ordering the target classes.On the other hand,we model the visual classes in a weakly supervised environment where a very limited number of training data is used. Meanwhile,we update both the attribute space and class models concurrently in an iterative algorithm.To do so,we establish a machine-machine interaction protocol that projects a Student(S)-Teacher(T)relationship that is available in the real world.

    1.1.Contributions

    We present three contributions in this paper.First,the attribute space and the class models are learned simultaneously in an iterative manner.This is very intuitive since the class models are constructed on the attribute space,and there must be a connection in a combined learning procedure.The second contribution is the bilateral learning paradigm where S updates the attribute space which,in fact,in fluences T's future attribute-based explanations. We can say that T is also learning or it should generate its responses with respect to S'inferences on the attributes.Also note that attributes are initialized randomly and T's attribute-based explanations are prepared online in an unsupervised manner.The last contribution is about the virtual growth of the training set.Unlike semi-supervised methods where each sample is used only once for the learning procedure,T does not reveal the true label of the selected image if the belief is false.So the number of training data is increased by dropping the falsely predicted sample back into the unlabeled pool.In the experiments,we see that the number of iterations is almost doubled,and this leads to a better performance in visual recognition.

    The general structure of the rest of the paper is as follows: Section 2 summarizes the related work about attribute-based applications with their important aspects.Then,we explain the proposed work in details in section 3,and the experimental evaluations are presented with comparative results on two popular datasets in section 4.We conclude the paper with our clear inferences in section 5.

    2.Related work

    The literature of attribute-based computer vision problems can be generally summarized in the types and extraction methods of attributes,the applications and datasets on which they are implemented,and evaluation criteria in the experiments,as shown in Fig.1.

    2.1.Attribute types

    The importance of an attribute for each class varies for recognition purposes.Thus,we may produce an image/categoryattribute co-occurrence matrix such that attribute membership for each class/image and the corresponding correlations with respect to the attributes may be captured.These relations can be produced in a supervised manner beforehand;or in a semisupervised/unsupervised fashion.Some researchers suggest that binary correlations would be suf ficient while others claim real valued ranking scores are essential to measure the attribute strength among categories/samples.

    Relative attributes are first introduced by Parikh and Grauman in Ref.[6]with the assumption that semantically more enriched data descriptions and discrimination will be achieved if we use relative class memberships on attributes,instead of binary relations.Biswas and Parikh propose an interactive learning scheme for both object and face recognition with relative attributes in Ref.[15].They improve this semi-supervised work in Ref.[10]by weighting negative candidate samples,selecting the query images and updating the attributes.

    Kovashka and Grauman[16]open a new perspective to learn binary attributes by modeling them with many visual de finitions instead of training only one discriminative model for each which is called shades of an attribute.Jayaraman and Grauman[17]study zero-shot classi fication paradigm with the assumption that traditional attribute scores are not suitable enough to predict them perfectly in unseen images.Chao-Yeh and Grauman[18]convey the zero-shot discussion on a new idea that class-sensitive attribute classi fiers can transfer the learning to infer the new ones unobserved during training.Liang and Grauman[19]explore active learning strategies to learn relative attributes better without exploiting all pairwise comparisons.

    Fig.1.Summary of the literature work on attribute-based computer vision problems.

    2.2.Attribute extraction

    In supervised settings,the attribute labeling can be constructed on whether the instances individually or the visual categories generally.Akata et al.[8]try to solve the attribute-based classi f ication problem by constructing a composite framework which embeds mapping images from both sides of low-level features and class labels into a common attribute space at once.Jing et al.[20] generate a deep Convolutional Neural Network(CNN)architecture to learn appearance and motion based binary attributes.Kiapour et al.[21]experiment a challenging task to extract deep learning features that are based on retrieval and similarity learning methods for this task.Xinlei and Abhinav[22]propose an approach to utilize web data for learning visual attributes through CNN.

    Unlike supervised learning where the attributes are named speci fically,one may not name the attributes but only use them to discriminate the data classes.Farhadi et al.[5]embed the object recognition problem into describing objects and mainly focus on attribute learning.Ma et al.[11]implement an algorithm to learn class-level relative attributes in an unsupervised manner where they produce relative attributes as many as the number of pairwise class combinations in a greedy fashion.Karayel and Arica[23] follow the similar way of[11,24]but they produce the attributes completely randomly and attribute selection is implemented afterwards.Yu et al.[25]formalize a category-attribute co-occurrence matrix for cross-category generalization.

    One may relate the attributes with each other such that hierarchical structure is established based on spatial layout,heritage and encapsulation aspects.Ran et al.[26]combine categoryspeci fic attributes and category-level information for generic instance search in the web.Victor et al.[27]extract relative attributes in different layers of the CNNs architecture by encoding the location,sparsity and the relationship between visual attributes. Local attributes,besides global ones,are also considered when the spatial ordering is important for recognition.But in this case,extra segmentation methods are required such as bounding boxes[4,7] or intensive labeling approaches[28]lead to additional cost to learn local attributes.

    3.Materials and methods

    3.1.Overview of the proposed method

    In this work,we introduce a new approach for incremental visual recognition to modelvisual categories simultaneously with the relative attributes in an incremental learning procedure.To do so, we establish a machine-machine interaction protocol that projects a Student(S)-Teacher(T)relationship available in the real world. Relative attributes are first initialized by T in a way that sorts the image classes randomly in an unsupervised manner.Because we do not aim to locate the attributes speci fically on spatial regions of the visual data,they are learned from global low-level features(i.e. GIST).The classes are then modeled by S with a limited number of category labeled data in the mid-level attribute space.Although it is T's initiative to generate random class orderings for the attribute learning,S then takes the responsibility to update both the attribute space and class models concurrently in an iterative algorithm. While S selects a sample,x(q),from a pool of unlabeled data each time and propagates it with the belief class,T paves the way for S' corrective actions on class modeling with its attribute-based, comparative and af firmative/rejective responses.

    Unlike[10,15]where the attribute-based explanations are prepared by the Amazon Mechanical Turk(AMT)workers before the learning procedure,the T machine decides online what explanation should be given to S for mentoring.While doing updates on class models,S also appends some candidate samples to its labeled dataset.The candidates are selected out of the unlabeled data,Uset, temporarily with respect to the conditions on the current attributes by constructing a weighting cube.This alternate routine which describes the attribute-class model updates continues until S runs out of the unlabeled data that leads to incremental learning for recognition.The flow chart of the algorithm is given in Fig.2.

    3.2.Preliminaries of initiative setup

    Relative attributes are used as a communication channel in bidirectional interactions between Student(S)and Teacher(T) machines.To do so,we first divide the training set randomly, apart from test set,randomly into three discrete subsets:Student(Sset),Teacher(Uset)and Validation (Vset)sets.Sset consists of a very limited number of class-labeled images per category,and S uses this subset to achieve initial class models.On the other hand,S does not know the class labels of Uset and it basically represents T's teaching materials about the task domain.S utilizes Uset for both picking a sample in its belief propagation through questioning and for selecting some candidates in a weighting scheme to update its current category models. On the contrary,T handles this subset for preparing the best attribute-based response to indicate why the belief is correct or not. Vset is used for optimizing the free parameters of the algorithm and the initial setup is summarized in Fig.3.

    Another important detail of the proposed work is to construct an attribute space which constitutes the communication channel between S and T machines.Given a class based ordering,which relates every category to each other with a less/more and similarity conditions,we use the Newton's method in Ref.[6]for a relative attribute as:

    where rmis the real valued ranking score for the training instance,xi,on the attribute basis,is the parameter vector of the relative attribute model.Omis the set which includes pairs of data instances holding for the more/less conditions whileSmrefers to the set of similar samples.So the optimum solution would then order the classes on the weight vector,wm,by minimizing the cost function in(4);where Z is the constant that regulates the balance between weight decreasing and the non-negative slack variables,εijandγij.This results in maximizing the margin between classes in the order de finition ofam.

    T initializes the relative attribute space on Sset by randomly generated class ordering[23]for each attribute in an unsupervised manner,and shares this information with S.After preparation of the subsets for training and random initialization of the attribute space by T,it is now time for S to initialize category models by using Sset.S uses RBF kernel-based one-vs-all SVM[29]for class discriminants and the free parameters are optimized in a grid search method with 10-fold cross validation on Vset.To summarize,low level features, x(i),are first projected into attribute space,A={am}∈?M,and classi fication is implemented in this mid-level with Sset at the beginning.

    Fig.2.Flow chart of the proposed algorithm.

    3.3.Student-teacher interactions for visual recognition

    Incremental learning corresponds to gradually updating the attribute space and category models alternately in S-T interactions until S labels all of the supplementary data set,Uset,correctly with the help of T's attribute-based explanations.In detail,S picks a sample,x(q),out of Uset which will be as much informative as possible for its class learning and this is achieved by maximum entropy,H,[10]:

    wherep(x(i))yis the probability output of SVM for the sample,x(i), for each class,y.Hence,the class label c(j)with the highest probability,y*,is attached to the query sample,x(q),and it is questioned for an explanation like‘Does x(q)belong to class c(j)?’As aforementioned,the falsely predicted sample is dropped back into Uset which helps increasing the training set virtually.

    When S machine often picks the same set of query samples,x(q), out of Uset,T machine would generate similar attribute-based explanations for the S'beliefs.Because S uses these explanations for the updates of attribute space,over fitting may occur respectively. To avoid over fitting at this point,we propose three enhancements for S.First,S machine also propagates its best prediction once in a while(i.e.one in ten iterations)on the contrary to the selection method which is based on the maximum entropy.Secondly,S restricts one sample image to be selected often by deploying a counter for each.For instance,if S sends the x(q)back to the Uset its counter will be initialized(for example set to three).The counters are then decreased by one through iterations,and the previously selected images are not allowed to be chosen again until their counters reach zero,i.e.not for the three next runs in this case. Finally,S keeps track of the selection history of samples in the current Uset.Hence,the image is not propagated again with a class label which has been predicted falsely in the previous iterations.

    After S propagates its belief in a query it is now T's turn to reply, and this process is displayed in Fig.4.The attributes are computed relatively with respect to the class labels and T already knows the class labels of both Sset and Uset.So one may append an attributebased explanation to the decision of’accepted’or’rejected’by selecting an attribute and its comparative term.Unlike semisupervised learning[30,31],S is in fluenced in the complementary attribute space for category modeling in a way that the query sample,x(q),is sent back to Uset if the prediction is false,or it is added to Sset,otherwise.T is then responsible to select the best attribute-condition pair for the explanation and this is achieved by first projecting x(q)and the datasets,Sset and Uset,on each attribute axes withra(x(i))=(x(i))Twa,wherera(x)is the ordering function computed as in Ref.[6]and wais the weight vector of attribute am.Thereafter,the samples are sorted increasingly withtheir ranks,ra,as in Ref.[15]and the voting method is implemented as:

    Fig.3.Preliminaries of the proposed incremental learning algorithm.

    where indnum(·)refers to the index number of samples on the attribute axes in the ordered list.The intuition is that if x(q)does not belong to the falsely predicted class,Cpredicted,but it belongs to another class,Ctrue,where Ctrue>Cpredictedon am,thenwe expect its vote to be positively high for that attribute.On the contrary,the vote should be as low as possible in its absolute form when the prediction is true.So T has 2 M options(i.e.more or less conditions for M attributes)to choose for false and M options for true beliefs since the condition is already selected as’similar’.

    Finally,T may mislead S by just using the probabilities computed in(8).For example,giving an explanation like‘No,x(q)is not as natural(am)as c(1)’contradicts the reality when it is the case‘Ctrue>c(1)on naturalness'.Because T knows the category orderings on attributes in addition to Ctrueof x(q),the response may be reestablished with the next highest probability.Also note that there is no need of such considerations when the belief is true,and the attribute with the lowest vote(i.e.unsigned)is selected with the condition of similarity.

    3.4.Updating parameters of the learning algorithm

    3.4.1.Attribute spacef ine-tuning

    Basically,the student-teacher interaction produces a pair of query-response through the iterations which triggers updating the parameters of attribute space and category models alternately. Assuming that T accepts S belief,x(q)is appended to Sset and subtracted from Uset since the answer reveals the ground truth category label of x(q).Thus,all attributes can be updated by creating new pairs with x(q)in its class orderings.The situation becomes interesting when T rejects the prediction because the true class label of x(q)is not given in the answer and it is not added to Sset.But S still finds additional pairs to update only the selected attribute by perceptual induction.For instance in‘No,x(q)is more natural than class c(1)’,it is known that x(q)is not in the predicted class with’more’condition on attribute’naturalness'.So one may simply produce new ordering pairssuch that query image visualizes attribute amstronger thanon Sset.

    Additionally,S may infer even more pairs by using its prior knowledge.When the class-based ordering on amisand the previous answer is given for false prediction,then x(q)is actually more natural thanThe same idea holds for the‘less’condition except reverse ordering rule.Hence,this would increase the number of pairing samples for attribute learning which leads to better discrimination.Also note that newly added pairs are deleted for the next iteration since S has not discovered the class label of x(q)yet,unlike in case of true prediction.Besides,previously updated attribute weight vectors, wa,are carried through iterations for incremental learning insteadof random initialization,and only the paring samples which violate the constraints of ranking margin are used in the attribute update procedure.It accounts for avoidance of the over fitting when using the same ordering pairs through iterations.

    Fig.4.Student-Teacher interaction scenario.

    3.4.2.Selecting candidates for updating the class models

    Like in attribute updating,S makes inferences out of T's answers but now it is for picking the candidate samples from Uset temporarily to update class models.To do so,S produces a Uset’Weighting Cube’on the Uset in which the weights are summed cumulatively in iterations for each candidate with respect to positive and negative sides,in addition to target categories.The weighting scheme is visualized in Fig.5.In detail,the candidate samples,Xcandidates={x(p)}p=1,2,…,U;x(p)∈Uset,are chosen on the attribute base,am,which has been given in T's explanation by first computing the ranks of all samples,ra(x(i))=(x(i))Twa; x(i)∈(Uset+Sset).Afterwards,the samples are sorted and positions of the Uset samples in the sorted list are checked with x(q)as it is the benchmark on the condition embedded in T's response.If S predicts x(q)correctly to be c(1),for instance,the explanatory condition will be’similarity’,and the candidates should be even more similar to c(1)than x(q)on the am.This is achieved by first finding the median image of c(1)(i.e.μ(x(j)),see Fig.5a.)and then marking the symmetric area which is centered on the median point and bounded by x(q).The Uset samples projected on this region are selected as the’positive’candidates for c(1)in this iteration only if they are not pinned as’negative’for c(1)previously.

    On the other hand,the‘negative’candidates for the predicted class can be achieved in false predictions easily in a similar manner except relative conditioning which determines the side moving away from x(q)(i.e.to the left for‘less’and to the right for’more’). The intuition for‘less’condition is that if x(q)is not as amas c(3),the samples which are even less amthan x(q)would not belong to c(3), either.The same idea holds for the’more’condition in the opposite direction(see Fig.5b and c).After selecting the candidate positive/ negative samples,x(p),for the predicted class,Cpredicted,S computes their cumulative weights as:

    where UB is the upper bound of the iterations andσis the scale factor.Summation yields to cumulative votes over iterations which is independent on the selected attribute,am.So the weights happen to be the distance that refers to the number of images between the candidate and x(q).Also note that the exponential factor in(9) ensures that weights are added with an increasing ratio.Eventually, S selects some candidate samples out of Uset with their weights in each iteration,and these candidates will be appended to Sset for updating the category models.

    3.4.3.Updating the class models

    So far,S fine-tunes the attribute space and picks some candidate samples with their weights from Uset which will be appended to Sset as a temporary extension.Now it is time to update the category models,and this is achieved by one-vs-allL2RBF kernel based SVM[29].In detail,the samples in the Sset are known with their labels by S,so their weights are naturally 1 like in Ref.[15].Hence,the candidate weights are first normalized to[0,1]such that they are summed to 1 for true predictions while the maximumvalue is set to 1 for false predictions because absolutely negative samples exist in the candidate list.

    Fig.5.Candidate selection with a weighting scheme.{x(1),x(2),x(3),x(4)}∈Uset are the selected candidates for an explanation of:a)x(q)is quite as amas c(1),b)x(q)is not as amas c(3),c)x(q)is more amthan c(4).

    Once S predicts x(q)truly the candidates are then labeled as Cpredictedwith their normalized weights.The situation becomes confusing when T rejects the belief because S simply assumes that the candidates are not be in Cpredicted,either.For example,if T answers‘x(q)is not as amas c(3)’and the class-ordering is c(2)>c(3)>c(4)>c(1)on am,S may expect that the candidates x(p),in addition to x(q),may come from c(4)or c(1).The decision is trivial when there is only one candidate category(i.e.c(1)if c(3)is interchanged with c(4))and the negative side of the Cpredictedvote is taken directly as the candidate weight.But S should make its best guess for x(p)if there are more possible classes.In detail,S picks the best one among candidate classes by using its current category models if there is no weighting for these classes yet.Otherwise,the one with the highest vote,computed by the difference between positive and negative weights of x(p)on the candidate class,is chosen directly.

    So long as S determines the weights and class labels for the candidate samples,x(p),which are appended to Sset temporarily, SVM algorithm is implemented on the new training set by initializing the model parameters with the previous ones in each iteration.The free parameters are optimized in a grid search method with 10-fold cross validation on Vset as mentioned in the’preliminaries(see 3.2)’subsection.Also note that class models and attribute space are not updated each time when S achieves worse accuracy performance than the previous setups on the Vset.The intuition is that T's explanations sometimes may not help S improve its visual recognition because of over fitting,and it can downgrade the performance in iterations.The alternate update procedure continues iteratively until S empties Uset with its correct predictions.

    4.Performance evaluation

    4.1.Experimental setup

    We use two real world datasets to evaluate the proposed algorithm for comparative experiments with the other attribute based methods:Outdoor Scene Recognition(OSR)[6,10,11]that contains 2688 images in 8 scene categories,and Public Figure Face(PUBFIG) [6,15,23]which has 800 images from 8 face classes.The global low level features(i.e.512D GIST and additional 30D LAB color histograms for PUBFIG)are also provided with the datasets.We repeat the experiments 30 times with different training/test splits,and the average results are noted for generalization.In each run,there exist 15 samples in the Sset,25 images in the Uset through random selection,and the rest is reserved for the test.Vset is generated from Sset+Uset by injecting Gaussian noise.Furthermore,the supervised attributes which are given with the datasets are also used(6 for OSR and 11 for PUBFIG[6])for the performance evaluation.In the implementation,LibSVM[29]algorithm is utilized for category modeling while Newton's method in Ref.[6]is used for the extraction of relative attributes.Also note that RBF kernel outputs are whitened as in Ref.[32]for normalization before SVM.

    Basically,the proposed work is compared with two framework methods of attribute-based recognition:Direct Attribute Prediction (DAP)and Semi-supervised Attribute Classi fication(SSAC).DAP offers a two-phase approach where the attribute space and class models are learned independently.Hence,Sset and Uset are now combined in a joined set.We first learn the relative attributes; thereafter categories are modeled on this mid-level via the joined set.SSAC,on the other side,represents’in-between’methodology between DAP and the proposed work.The attribute space is generated with the joined set,Sset+Uset,and the attribute weight vectors,w,are fixed during the class modeling like in DAP.S initializes the classi fiers on the attribute space by using Sset at the beginning,and it picks a sample out of Uset with its current best guess in an iterative manner.So,one sample is transferred fromUset to Sset with the belief class label and S updates the class models with the new Sset in each run.

    4.2.Multi-class classification

    Our main concerns in multi-class classi fication are whether:a. Attributes constitute a fruitful intermediate space for data representation,b.Unsupervised attributes are more discriminative than the crowd sourced attribute labeling in supervision,c.Learning attributes and category models incrementally in a machinemachine interaction increases the recognition accuracy when compared with DAP and SSAC frameworks.

    The plot of mean accuracy vs.varying number of attributes for OSR and PUBFIG datasets are displayed in Fig.6.With regards to the supervised attributes alone,all three algorithms seem to achieve the similar performance despite the fact that the proposed and SSAC algorithms maintain a better start.We believe that it is due to the limited number of attributes and the class orderings on attributes are akin to each other.We learn only 28 unsupervised attributes in order to compare the proposed work with the other studies in literature.For these,all methods achieve much lover accuracy at the beginning but they make up the difference quickly when compared with the supervised attributes.As expected,the performance increases little about 1-2%for DAP and 3-4%for SSAC overall,whereas an obvious rise is experienced with the proposed algorithm,that is up to 6-8%for the 28 unsupervised attributes. The same evaluation is still true if we combine both.When the unsupervised attributes are appended,the accuracy rises up again about 3%with respect to the single usage of unsupervised attributes.

    As SSAC and the proposed method are common in learning the class models iteratively,we display their plots of accuracy vs.iteration step in Fig.7.Remembering that the relative attributes are learned with Sset+Uset initially,SSAC keeps them fixed during the class model updates whereas S machine updates the attribute space and classi fiers concurrently with respect to T's attribute-based explanations in the proposed work.Also note that the query image, x(q),is transferred with its belief class label directly into the Sset each time in the SSAC algorithm.So the iteration bound is restricted simply to the number of samples in the Uset,i.e.200 for 25 images per category.On the contrary,the bound varies in the proposed method since the query image is dropped back into the Uset when it is predicted falsely.

    In Fig.7,it is observed that the SSAC makes a better startof about 2-4%than the proposed method.Because SSAC uses the joined set, Sset+Uset,for the attribute space initialization,we assume that SSAC has started to learn the attributes more consistently with a larger dataset,as expected.But the accuracy trend of SSAC is not as steep as the one of the proposed work during further iterations.For example in the PUBFIG experiments,the best accuracy level of the SSAC is already captured at about iteration#230 and 4%final increase is achieved in the proposed algorithm.We conclude that theperformance is increased by expanding the training set virtually while some Uset samples are used multiple times in iterations.

    Fig.6.Accuracy performance analysis with varying number of attributes on:a)OSR,b)PUBFIG.

    Fig.7.Iterative performance analysis of SSAC and the proposed work on OSR and PUBFIG datasets.

    Receiver Operating Characteristic(ROC)curve is frequently used in literature to evaluate the performance of classi fiers.Basically,the ratio of false and true positive samples is plotted by changing thresholds in a step-wise manner.The classi fier is regarded as more successful when its plot rises up earlier and sharper than the others.We compare the performances of the DAP,SSAC and proposed work with supervised and combined(i.e. supervised+unsupervised)attributes on ROC curves for both datasets in Fig.8.As seen,the accuracy is increased obviously when all attributes are used together,and this con firms that theunsupervised attributes add discriminative power in dimensionality.Additionally,the proposed algorithm outperforms the others clearly while SSAC is slightly better than the DAP method.

    Fig.8.ROC analysis of the DAP,SSAC and the proposed algorithms on:a)OSR,b)PUBFIG.

    Table 1 Performance(%)comparison of the algorithms.

    Finally,the proposed method is compared with the similar approaches in literature on the same experimental setup,and the mean±std accuracy results of 30 repeats are listed in Table 1.BINs, PCA and FLD algorithms are actually used for dimension reduction. Nevertheless,the basis vectors which are extracted during their implementations help representing the data in a new feature space, so they are included as baselines for this reason.The other methods generate supervised/unsupervised attributes in the mid-level for visual recognition,like the proposed work.The results in Table 1 show that the proposed method outperforms the other approaches for about minimum of 2%although the attributes are produced by random class orderings unlike in Ref.[11],and there is no speci fic attribute selection,differently from Ref.[23].The intuition behind such a notable result is that we shape the attribute space simultaneously with the class models and the number of training samples is increased during attribute-based feedbacks.We also assume that the expanded number of unsupervised attributes with distinct class orderings establish a better representation without human laboring,leading to more effective classi fiers.

    5.Conclusion

    In this study,we use attributes as the mid-level representation, instead of low-level features,in an incremental learning procedure for visual recognition.Attributes constitute a communication channelthrough S-T machine interactions.The conversation mainly indicates a question-answer protocol where S selects an informative sample from the unlabeled pool of data,Uset,each time and T gives attribute-based explanations for mentoring.Also note that the training set is expanded virtually since the query image is dropped back into the Uset without exposing its ground-truth category label when S predicts it falsely.So we introduce a student centered learning method where S is actually responsible for updating the attribute space and class models simultaneously in an iterative manner.Although T initiates the relative attributes by sorting the image categories randomly in an unsupervised fashion, S updates the attributes with the selected query images from then on.This means that T also learns in time how the student interprets the representative space on which the class discriminants are built. Additionally,S does not use only its labeled dataset,Sset,for updating the class models,but it also implements a candidate selection method to append some samples from Uset to Sset in a weighting scheme.

    We experiment the proposed work comparatively with the frameworks DAP and SSAC,and other similar approaches in literature on two popular mid-scale datasets,OSR and PUBFIG.Because classi fication is the focal point,the mean accuracy vs.iteration step/ number of attributes and the ROC curves are the main themes for analysis during multiple experiments.To summarize,the proposed method outperforms other works signi ficantly,and it is concluded that the unsupervised attributes exploit the recognition performance when learned incrementally and simultaneously with the class models even though they are achieved by random class orderings without human laboring.

    Acknowledgements

    This work was supported in part by the Scienti fic and Technological Research Council of Turkey(TUBITAK)2214B.14.2 TBT.0.06.01.214.83.

    Appendix A.Supplementary data

    Supplementary data related to this article can be found at http:// dx.doi.org/10.1016/j.trit.2017.01.001.

    [1]V.Ferrari and A.Zisserman,Learning visual attributes,Neural Information Processing Systems Conference,3-8 December 2007,Vancouver,Canada,pp. 433-440.

    [2]C.H.Lampert,H.Nickisch,S.Harmeling,Attribute-based classi fication for zeroshot visual object categorization,J.IEEE Trans.Pat.Anal.Mach.Intel.36(2014) 453-465.

    [3]C.H.Lampert,H.Nickisch and Harmeling S,Learning to detect unseen object classes by between-class attribute transfer,IEEE CVPR Conference,20-25 June 2009,Miami,FL,USA,pp.951-958.

    [4]A.Farhadi,I.Endres and D.Hoiem,Attribute-centric recognition for crosscategory generalization,IEEE CVPR Conference,13-18 June 2010,San Francisco,CA,USA,pp.2352-2359.

    [5]A.Farhadi,I.Endres,D.Hoiem and D.Forsyth,Describing objects by their attributes,IEEE CVPR Conference;20-25 June 2009,Miami,FL,USA,pp. 1778-1785.

    [6]D.Parikh and K.Grauman,Relative attributes,IEEE ICCV Conference,6-13 November 2011,Barcelona,Spain,pp.503-510.

    [7]G.Sharma,F.Jurie and C.Schmid,Expanded parts model for human attribute and action recognition in still images,IEEE CVPR Conference,23-28 June 2013,Portland,OR,USA,pp.652-659.

    [8]Z.Akata,F.Perronnin,Z.Harchaoui and C.Schmid,Label embedding for attribute based classi fication,IEEE CVPR Conference,23-28 June 2013,Portland,OR,USA,pp.819-826.

    [9]A.Bosch,M.Xavier,R.Marti,A review:which is the best way to organize/ classify images by content,J.Image Vis.Comput.25(2007)778-791.

    [10]A.Parkash and D.Parikh,Attributes for classi fier feedback,Springer ECCV Conference,7-13 October 2012,Florence,Italy,pp.354-368.

    [11]S.Ma,S.Sclaroff and N.I.Cinbis,Unsupervised learning of discriminative relative visual attributes,Springer ECCV Conference,7-13 October 2012, Florence,Italy,pp.61-70.

    [12]Y.Wang and G.Mori,A discriminative latent model of object classes and attributes,Springer ECCV Conference,5-11 September 2010,Crete,Greece,pp. 155-168.

    [13]C.Wah and S.Belongie,Attribute-based detection of unfamiliar classes with humans in the loop,IEEE CVPR Conference,23-28 June 2013,Portland,OR, USA,pp.779-786.

    [14]M.Rastegari,D.Parikh and A.Farhadi,Multi-attribute queries:to merge or not to merge,IEEE CVPR Conference,23-28 June 2013,Portland,OR,USA,pp. 3310-3317.

    [15]A.Biswas and D.Parikh,Simultaneous active learning of classi fiers and attributes via relative feedback,IEEE CVPR Conference,23-28 June 2013, Portland,OR,USA,pp.644-651.

    [16]A.Kovashka,K.Grauman,Discovering attribute shades of meaning with the crowd,Int.J.Comput.Vis.114(2015)56-73.

    [17]D.Jayaraman and K.Grauman,Zero-shot recognition with unreliable attributes,Neural Information Processing Systems Conference,8-13 December 2014,Montreal,Canada,pp.3464-3472.

    [18]C.Chao-Yeh and K.Grauman,Inferring analogous attributes,IEEE CVPR Conference,23-28 June 2014,Columbus,OH,USA,pp.200-207.

    [19]L.Liang and K.Grauman,Beyond comparing image pairs:setwise active learning for relative attributes,IEEE CVPR Conference,23-28 June 2014,Columbus,OH,USA,pp.208-215.

    [20]S.Jing,K.Kai,C.L.Chen,and W.Xiaogang,Deeply learned attributes for crowded scene understanding,IEEE CVPR Conference,7-12 June 2015,Boston,MA,USA,pp.4657-4666.

    [21]M.H.Kiapour,H.Xufeng,L.Svetlana,C.B.Alexander and L.B.Tamara,Where tobuy it:matching street clothing photos in online Shops,IEEE ICCV Conference, 7-13 December 2015,Santiago,CA,USA,pp.3343-3351.

    [22]C.Xinlei and G.Abhinav,Webly supervised learning of convolutional networks,IEEE ICCV Conference,7-13 December 2015,Santiago,CA,USA,pp. 1431-1439.

    [23]M.Karayel and N.Arica,Random attributes for image classi fication,IEEE SIU Conference,24-26 April 2013,Girne,TRNC,pp.1-4.

    [24]N.Kumar,A.C.Berg,P.N.Belhumeur and S.K.Nayar,Attribute and simile classi fiers for face veri fication,IEEE ICCV Conference,29 September-2 October 2009,Kyoto,Japan,pp.365-372.

    [25]F.X.Yu,L.Cao,R.S.Feris,J.R.Smith and S.Chang,Designing category-level attributes for discriminative visual recognition,IEEE CVPR Conference, 23-28 June 2013,Portland,OR,USA,pp.771-778..

    [26]T.Ran,W.M.S.Arnold and C.Shih-Fu,Attributes and categories for generic instance search from one example,IEEE CVPR Conference,7-12 June 2015, Boston,MA,USA,pp.177-186.

    [27]E.Victor,C.N.Juan and G.Bernard,On the relationship between visual attributes and convolutional networks,IEEE CVPR Conference,7-12 June 2015, Boston,MA,USA,pp.1256-1264.

    [28]S.Zhiyuan,Y.Yongxin,M.H.Timothy and X.Tao,Weakly supervised learning of objects,attributes and their associations,Springer ECCV Conference,6-12 September 2014,Zurich,Switzerland,pp.472-487.

    [29]C.C.Chang,C.J.Lin,LIBSVM:a library for support vector machines,ACM T Intel.Syst.Tech.27(2011)1-27.

    [30]J.Choi,M.Rastegari,A.Farhadi and L.S.Davis,Adding unlabeled samples to categories by learned attributes,IEEE CVPR Conference,23-28 June 2013, Portland,OR,USA,pp.875-882.

    [31]A.Shrivastava,S.Singh and A.Gupta,Constrained semi-supervised learning using attributes and comparative attributes,Springer ECCV Conference,7-13 October 2012,Florence,Italy,pp.369-383..

    [32]A.Coates,Y.N.G.Andrew,H.Lee,An analysis of single-layer networks in unsupervised feature learning,J.Mach.Learn Res.15(2011)215-223.

    [33]E.Ergul,S.Erturk and N.Arica,Unsupervised relative attribute extraction,IEEE SIU Conference,24-26 April 2013,Girne,TRNC,pp.1-4.

    [34]E.Ergul,M.Karayel,O.Timus,E.Kiyak,Unsupervised feature learning for midlevel data representation,J.Nav.Sci.Eng.12(2016)51-79.

    Emrah Ergulreceived the B.S.degree in Electrical and Electronics Engineering from Turkish Naval Academy, Istanbul,Turkey,in 2001.From 2001 to 2007,he served in the Navy Fleet Command,Kocaeli,Turkey as a weapon systems and electronics of ficer.He received the M.S.degree in Computer Engineering from Naval Science and Engineering Institute,Istanbul,Turkey,in 2009.He also studied at the Computer Science Dept.of University of Shef field,UK for one semester with respect to Erasmus Student Exchange Program while pursuing the M.S.degree.He studied in the Computer Vision and Robotics Laboratory at the Beckman Institute of the University of Illinois,Urbana-Champaign IL,USA,between 2013 and 2014 with a scholarship granted by The Scienti fic and Technological Research Council of Turkey.He received the Ph.D.degree in Electronics and Communication Engineering from Kocaeli University,Kocaeli,Turkey in 2016.He has been working for Navy Inventory Control Center,Kocaeli,Turkey since 2009 for stock estimation and process control.His research interest includes computer vision for image retrieval and classi fication,logistics decision support systems and data exchange standards for inventory planning,pattern recognition analysis,data mining and machine learning applications for the recommender systems.

    E-mail addresses:106103002@kocaeli.edu.tr,emergul13@yahoo.com.

    Peer review under responsibility of Chongqing University of Technology.

    http://dx.doi.org/10.1016/j.trit.2017.01.001

    2468-2322/?2017 Chongqing University of Technology.Production and hosting by Elsevier B.V.This is an open access article under the CC BY-NC-ND license(http:// creativecommons.org/licenses/by-nc-nd/4.0/).

    Incremental learning

    Relative attribute

    Visual recognition

    大型黄色视频在线免费观看| 别揉我奶头 嗯啊视频| 天堂av国产一区二区熟女人妻| 在线观看66精品国产| 午夜福利18| 国国产精品蜜臀av免费| 久久九九热精品免费| 亚洲国产精品久久男人天堂| 波野结衣二区三区在线| 精品国产三级普通话版| 色精品久久人妻99蜜桃| 九九爱精品视频在线观看| 国产精品99久久久久久久久| 成人三级黄色视频| 91久久精品国产一区二区三区| 国产亚洲av嫩草精品影院| 亚洲国产精品久久男人天堂| 又紧又爽又黄一区二区| 亚洲av.av天堂| 99久久无色码亚洲精品果冻| 99热这里只有精品一区| 搞女人的毛片| 3wmmmm亚洲av在线观看| 欧美成人a在线观看| 一级av片app| 韩国av一区二区三区四区| 亚洲专区中文字幕在线| 琪琪午夜伦伦电影理论片6080| 国产淫片久久久久久久久| 老师上课跳d突然被开到最大视频| 亚洲av五月六月丁香网| 日本 av在线| 91在线精品国自产拍蜜月| 亚洲欧美激情综合另类| 在线国产一区二区在线| 久久热精品热| 99九九线精品视频在线观看视频| 国产69精品久久久久777片| 午夜免费男女啪啪视频观看 | 国产精品av视频在线免费观看| 观看美女的网站| 在线看三级毛片| 亚洲av中文字字幕乱码综合| 久久久久久九九精品二区国产| 精品午夜福利在线看| 久久久久九九精品影院| 色尼玛亚洲综合影院| 成人高潮视频无遮挡免费网站| 精品久久久久久,| 黄色视频,在线免费观看| 一进一出抽搐动态| 美女被艹到高潮喷水动态| 欧美日韩黄片免| 免费不卡的大黄色大毛片视频在线观看 | 岛国在线免费视频观看| 99久国产av精品| 亚洲精华国产精华液的使用体验 | 国产亚洲精品久久久久久毛片| 国产伦精品一区二区三区视频9| 69av精品久久久久久| 国产v大片淫在线免费观看| 国产伦在线观看视频一区| 最好的美女福利视频网| 观看美女的网站| 亚洲精品456在线播放app | 最后的刺客免费高清国语| 直男gayav资源| 一个人看的www免费观看视频| eeuss影院久久| 一区二区三区四区激情视频 | 非洲黑人性xxxx精品又粗又长| 99热精品在线国产| 久久人人精品亚洲av| 99九九线精品视频在线观看视频| 久久精品国产99精品国产亚洲性色| 亚洲无线观看免费| 人人妻人人澡欧美一区二区| 女人十人毛片免费观看3o分钟| 老司机福利观看| 欧美精品啪啪一区二区三区| 我的女老师完整版在线观看| 在线免费十八禁| 日韩一区二区视频免费看| 欧美日韩乱码在线| 中出人妻视频一区二区| 日本熟妇午夜| 久久精品国产自在天天线| 91在线精品国自产拍蜜月| 欧美日韩亚洲国产一区二区在线观看| 99视频精品全部免费 在线| 99久久中文字幕三级久久日本| a在线观看视频网站| 最近最新中文字幕大全电影3| h日本视频在线播放| 九九久久精品国产亚洲av麻豆| 国产精品久久久久久av不卡| 中文字幕久久专区| 又粗又爽又猛毛片免费看| a级毛片a级免费在线| 久久精品国产亚洲网站| 色综合色国产| 亚洲18禁久久av| 亚洲成人久久爱视频| 国产精品人妻久久久久久| 国产男靠女视频免费网站| 亚洲最大成人av| 最新在线观看一区二区三区| 国产老妇女一区| 久久久久久伊人网av| 男插女下体视频免费在线播放| 中文字幕久久专区| 男女啪啪激烈高潮av片| 免费观看人在逋| 久久草成人影院| 欧美绝顶高潮抽搐喷水| 99热只有精品国产| 国产人妻一区二区三区在| 色综合婷婷激情| 国产精品福利在线免费观看| 国模一区二区三区四区视频| 国产激情偷乱视频一区二区| 亚洲精品成人久久久久久| 男女视频在线观看网站免费| 欧美潮喷喷水| 亚洲精品一区av在线观看| 校园人妻丝袜中文字幕| 99久久成人亚洲精品观看| 亚洲第一区二区三区不卡| 久久久久久国产a免费观看| 少妇人妻精品综合一区二区 | 成人二区视频| 国产一级毛片七仙女欲春2| 国产黄a三级三级三级人| 97超级碰碰碰精品色视频在线观看| 搞女人的毛片| 十八禁网站免费在线| 亚洲人与动物交配视频| 欧美不卡视频在线免费观看| 亚洲熟妇中文字幕五十中出| 亚洲成a人片在线一区二区| 亚洲va日本ⅴa欧美va伊人久久| 国产主播在线观看一区二区| 99久久精品国产国产毛片| 国产欧美日韩精品一区二区| 成人国产麻豆网| 亚洲欧美清纯卡通| 国产伦精品一区二区三区四那| 亚洲成人精品中文字幕电影| 成人无遮挡网站| 中文字幕久久专区| avwww免费| 日本一二三区视频观看| 中文字幕av在线有码专区| 午夜福利18| 嫩草影院入口| 韩国av一区二区三区四区| 久久亚洲精品不卡| 桃红色精品国产亚洲av| 国产探花极品一区二区| 国产精品国产三级国产av玫瑰| 简卡轻食公司| 国内精品一区二区在线观看| 亚洲内射少妇av| 国产亚洲欧美98| h日本视频在线播放| 免费高清视频大片| 欧美成人免费av一区二区三区| 一区二区三区激情视频| АⅤ资源中文在线天堂| 中文字幕免费在线视频6| 日本-黄色视频高清免费观看| 久久精品综合一区二区三区| 色尼玛亚洲综合影院| 99久久久亚洲精品蜜臀av| 国产黄色小视频在线观看| 99久久成人亚洲精品观看| 久久热精品热| 色播亚洲综合网| 亚洲在线自拍视频| av天堂在线播放| 99久久无色码亚洲精品果冻| 联通29元200g的流量卡| 男人舔女人下体高潮全视频| 亚洲人与动物交配视频| 日韩中文字幕欧美一区二区| 国产精品女同一区二区软件 | 淫秽高清视频在线观看| 国产综合懂色| 久久精品国产自在天天线| 在现免费观看毛片| 久久精品国产亚洲av涩爱 | 欧美bdsm另类| 狂野欧美白嫩少妇大欣赏| 老熟妇乱子伦视频在线观看| 亚洲精品国产成人久久av| 成人亚洲精品av一区二区| 免费在线观看影片大全网站| 少妇猛男粗大的猛烈进出视频 | 久久久久精品国产欧美久久久| 久久热精品热| 欧美黑人巨大hd| 国产激情偷乱视频一区二区| 最新中文字幕久久久久| 韩国av在线不卡| 成年版毛片免费区| 最近视频中文字幕2019在线8| 亚洲第一电影网av| 97碰自拍视频| 给我免费播放毛片高清在线观看| 国产白丝娇喘喷水9色精品| 久久人人爽人人爽人人片va| 日本三级黄在线观看| 午夜福利高清视频| 最后的刺客免费高清国语| netflix在线观看网站| 国产精品av视频在线免费观看| 国产成人aa在线观看| 成人国产综合亚洲| 亚洲国产欧美人成| 国产又黄又爽又无遮挡在线| 18禁黄网站禁片免费观看直播| 最后的刺客免费高清国语| 搡老熟女国产l中国老女人| 免费av毛片视频| 国产精品一区www在线观看 | 午夜激情福利司机影院| 一级黄色大片毛片| 免费观看人在逋| 免费av观看视频| 久久精品国产清高在天天线| 又爽又黄无遮挡网站| 日本-黄色视频高清免费观看| 午夜亚洲福利在线播放| 精品久久久久久久人妻蜜臀av| 久久久成人免费电影| 午夜激情福利司机影院| 两个人的视频大全免费| 一个人观看的视频www高清免费观看| 国产av麻豆久久久久久久| 国产精品av视频在线免费观看| 国产中年淑女户外野战色| 亚洲av二区三区四区| 噜噜噜噜噜久久久久久91| 级片在线观看| 国产高清激情床上av| 三级毛片av免费| 免费av不卡在线播放| 天堂√8在线中文| 内地一区二区视频在线| 亚洲专区中文字幕在线| 真人一进一出gif抽搐免费| 国产精品国产高清国产av| 变态另类丝袜制服| 精品人妻一区二区三区麻豆 | 99久久中文字幕三级久久日本| 女人被狂操c到高潮| 99久久久亚洲精品蜜臀av| 夜夜爽天天搞| 老司机深夜福利视频在线观看| 美女高潮喷水抽搐中文字幕| 日本在线视频免费播放| 午夜福利欧美成人| 99热这里只有是精品在线观看| 永久网站在线| 97热精品久久久久久| 国产伦在线观看视频一区| 哪里可以看免费的av片| 亚洲人成网站在线播| 男人舔奶头视频| 久久九九热精品免费| 亚洲色图av天堂| 国产精品精品国产色婷婷| 国产在线精品亚洲第一网站| 免费av毛片视频| 成年女人毛片免费观看观看9| 悠悠久久av| 午夜老司机福利剧场| 熟妇人妻久久中文字幕3abv| 国产精品爽爽va在线观看网站| 亚洲av成人精品一区久久| 日韩中字成人| 干丝袜人妻中文字幕| 国产女主播在线喷水免费视频网站 | 国产成人影院久久av| 联通29元200g的流量卡| 日韩欧美国产一区二区入口| 久久精品综合一区二区三区| 热99re8久久精品国产| 免费看光身美女| 一a级毛片在线观看| 一区二区三区免费毛片| 十八禁国产超污无遮挡网站| 中文亚洲av片在线观看爽| 2021天堂中文幕一二区在线观| 亚洲国产欧洲综合997久久,| 能在线免费观看的黄片| 日日摸夜夜添夜夜添av毛片 | 亚洲最大成人手机在线| 亚洲av成人精品一区久久| 国产爱豆传媒在线观看| 欧美一区二区国产精品久久精品| 人妻少妇偷人精品九色| 麻豆一二三区av精品| 国产精品久久久久久亚洲av鲁大| 舔av片在线| 蜜桃久久精品国产亚洲av| 国产一级毛片七仙女欲春2| 我要搜黄色片| 亚洲天堂国产精品一区在线| 亚洲精品亚洲一区二区| 亚洲性夜色夜夜综合| 亚洲男人的天堂狠狠| 国产午夜福利久久久久久| 精品久久久久久久久av| 美女 人体艺术 gogo| 内射极品少妇av片p| 午夜福利欧美成人| 精品久久久久久久末码| 久久国产精品人妻蜜桃| 免费在线观看影片大全网站| 中文字幕熟女人妻在线| 欧美中文日本在线观看视频| 国内精品一区二区在线观看| 在线观看一区二区三区| 女人十人毛片免费观看3o分钟| 一本精品99久久精品77| 中文字幕av成人在线电影| 亚洲欧美日韩卡通动漫| 亚洲在线自拍视频| 国产三级在线视频| 国产精品久久久久久av不卡| 成人美女网站在线观看视频| 午夜福利在线在线| 丝袜美腿在线中文| av专区在线播放| 美女免费视频网站| 久久久久九九精品影院| 欧美+日韩+精品| 免费电影在线观看免费观看| aaaaa片日本免费| 看十八女毛片水多多多| 男女之事视频高清在线观看| 人妻制服诱惑在线中文字幕| 成人午夜高清在线视频| 看免费成人av毛片| 啪啪无遮挡十八禁网站| 精品福利观看| 日韩高清综合在线| 亚洲人成网站在线播放欧美日韩| 国模一区二区三区四区视频| 一区二区三区激情视频| 毛片女人毛片| 欧美日韩黄片免| 日本欧美国产在线视频| 日本黄大片高清| 亚洲最大成人手机在线| 国产女主播在线喷水免费视频网站 | 一边摸一边抽搐一进一小说| 日韩中文字幕欧美一区二区| 1024手机看黄色片| 午夜福利高清视频| 亚洲欧美激情综合另类| 深夜a级毛片| 桃色一区二区三区在线观看| 午夜福利高清视频| aaaaa片日本免费| 日本五十路高清| 精品无人区乱码1区二区| 中文字幕av成人在线电影| 熟女人妻精品中文字幕| 波多野结衣高清无吗| bbb黄色大片| 国产爱豆传媒在线观看| 久久久久性生活片| 1000部很黄的大片| 热99re8久久精品国产| 日韩欧美国产一区二区入口| 亚洲人成伊人成综合网2020| 欧美日本亚洲视频在线播放| 特大巨黑吊av在线直播| 日韩精品有码人妻一区| 高清在线国产一区| 日本精品一区二区三区蜜桃| 联通29元200g的流量卡| 有码 亚洲区| 少妇的逼水好多| 久久久久久国产a免费观看| 中文字幕高清在线视频| 日韩欧美免费精品| 如何舔出高潮| 成人国产麻豆网| 国内久久婷婷六月综合欲色啪| 久久这里只有精品中国| 国产高清三级在线| 有码 亚洲区| 少妇的逼水好多| 黄色视频,在线免费观看| 成人永久免费在线观看视频| 日韩av在线大香蕉| 变态另类成人亚洲欧美熟女| 成年人黄色毛片网站| 永久网站在线| 三级毛片av免费| 国产男靠女视频免费网站| 欧美3d第一页| 最新中文字幕久久久久| 久久国产精品人妻蜜桃| 国产男人的电影天堂91| 国产精品野战在线观看| 我要搜黄色片| 婷婷精品国产亚洲av| 一本久久中文字幕| 悠悠久久av| 国产成人av教育| 神马国产精品三级电影在线观看| 又爽又黄无遮挡网站| 国产一区二区在线观看日韩| 亚洲天堂国产精品一区在线| 亚洲精品粉嫩美女一区| 91在线观看av| 两个人视频免费观看高清| 黄色日韩在线| 亚洲欧美日韩高清专用| 此物有八面人人有两片| 特大巨黑吊av在线直播| 免费电影在线观看免费观看| 99久久精品一区二区三区| 久久久精品欧美日韩精品| 老司机深夜福利视频在线观看| 色播亚洲综合网| 国产熟女欧美一区二区| 久99久视频精品免费| 午夜福利高清视频| 亚洲精品456在线播放app | 午夜激情欧美在线| av中文乱码字幕在线| 欧美丝袜亚洲另类 | 精品欧美国产一区二区三| 悠悠久久av| 一级黄片播放器| 国产 一区精品| 亚洲图色成人| 亚洲内射少妇av| 搡老熟女国产l中国老女人| 亚洲欧美日韩卡通动漫| 亚洲国产高清在线一区二区三| 别揉我奶头~嗯~啊~动态视频| 婷婷精品国产亚洲av在线| 久久精品国产99精品国产亚洲性色| 日韩精品有码人妻一区| 黄色欧美视频在线观看| 欧美性猛交黑人性爽| 九九爱精品视频在线观看| 午夜福利在线在线| 久久久色成人| 国产乱人伦免费视频| 亚洲性夜色夜夜综合| 中文字幕久久专区| 亚洲av免费在线观看| 亚洲美女搞黄在线观看 | 午夜福利欧美成人| 一级黄片播放器| 我的老师免费观看完整版| 99国产精品一区二区蜜桃av| 中文字幕av在线有码专区| 88av欧美| 亚洲精品乱码久久久v下载方式| 成人毛片a级毛片在线播放| 99久久精品国产国产毛片| 一级黄色大片毛片| 91久久精品国产一区二区成人| 人人妻人人看人人澡| 国产精品人妻久久久影院| 日韩欧美在线二视频| 成人三级黄色视频| 午夜爱爱视频在线播放| 我的老师免费观看完整版| 午夜精品在线福利| 熟女电影av网| www.www免费av| 国产高清不卡午夜福利| 嫩草影院入口| 国产精品1区2区在线观看.| 亚洲国产精品成人综合色| 一个人看视频在线观看www免费| 在线观看美女被高潮喷水网站| 人人妻人人看人人澡| 亚洲av日韩精品久久久久久密| 99热精品在线国产| 亚洲国产欧洲综合997久久,| 国产av一区在线观看免费| 日韩精品有码人妻一区| 国产精品伦人一区二区| 国产午夜精品久久久久久一区二区三区 | 99久久久亚洲精品蜜臀av| 亚洲四区av| 三级男女做爰猛烈吃奶摸视频| 久久6这里有精品| 免费av观看视频| 丰满乱子伦码专区| 老司机深夜福利视频在线观看| 99riav亚洲国产免费| 成人国产一区最新在线观看| 欧美高清性xxxxhd video| 波多野结衣高清作品| 国产伦一二天堂av在线观看| 成人无遮挡网站| 欧美+亚洲+日韩+国产| 精品99又大又爽又粗少妇毛片 | 国产免费男女视频| 俺也久久电影网| 国产三级中文精品| 九九爱精品视频在线观看| 国产黄a三级三级三级人| 亚洲最大成人中文| 国产欧美日韩一区二区精品| 亚洲欧美日韩无卡精品| 久久人妻av系列| 狂野欧美激情性xxxx在线观看| 老熟妇乱子伦视频在线观看| 久99久视频精品免费| 国产不卡一卡二| 黄色欧美视频在线观看| 日本a在线网址| av.在线天堂| 欧美3d第一页| 国产精品人妻久久久影院| 国产免费男女视频| 男人和女人高潮做爰伦理| 人人妻,人人澡人人爽秒播| 尤物成人国产欧美一区二区三区| 亚洲狠狠婷婷综合久久图片| 女同久久另类99精品国产91| 麻豆精品久久久久久蜜桃| 九九久久精品国产亚洲av麻豆| 小蜜桃在线观看免费完整版高清| 精品久久久久久久末码| 亚洲精品在线观看二区| 乱码一卡2卡4卡精品| av在线天堂中文字幕| 国产老妇女一区| 两人在一起打扑克的视频| 国产伦精品一区二区三区视频9| 赤兔流量卡办理| 啪啪无遮挡十八禁网站| 91午夜精品亚洲一区二区三区 | 国产精品日韩av在线免费观看| 成年版毛片免费区| 在线免费观看不下载黄p国产 | 大又大粗又爽又黄少妇毛片口| 欧美成人性av电影在线观看| 人妻夜夜爽99麻豆av| 色av中文字幕| 国产成人福利小说| 婷婷精品国产亚洲av| 一级黄片播放器| 国产精品免费一区二区三区在线| 日日摸夜夜添夜夜添av毛片 | 亚洲真实伦在线观看| av黄色大香蕉| 黄色视频,在线免费观看| av.在线天堂| 如何舔出高潮| 琪琪午夜伦伦电影理论片6080| 亚洲精品一卡2卡三卡4卡5卡| 亚洲人成伊人成综合网2020| 国产精品国产三级国产av玫瑰| 国产探花极品一区二区| 国产真实乱freesex| 欧美bdsm另类| 日韩欧美在线乱码| 欧美成人性av电影在线观看| 变态另类成人亚洲欧美熟女| 国产欧美日韩精品亚洲av| 国产成人影院久久av| 国产精品久久久久久久电影| 国产人妻一区二区三区在| 亚洲乱码一区二区免费版| 精品午夜福利视频在线观看一区| 亚洲美女视频黄频| 九色国产91popny在线| 亚洲内射少妇av| 国产免费男女视频| 色精品久久人妻99蜜桃| 亚洲电影在线观看av| 男人和女人高潮做爰伦理| 久久欧美精品欧美久久欧美| 欧美3d第一页| 国产伦精品一区二区三区视频9| 观看免费一级毛片| 欧美一区二区国产精品久久精品| 欧美人与善性xxx| 在线免费十八禁| 欧美日本亚洲视频在线播放| 亚洲avbb在线观看| 亚州av有码| 国产伦精品一区二区三区四那| 在线观看美女被高潮喷水网站| 黄色日韩在线| 国产真实乱freesex| 国产精品无大码| 久久久久精品国产欧美久久久| 国产国拍精品亚洲av在线观看| 成人精品一区二区免费| 一级毛片久久久久久久久女| 国产毛片a区久久久久| 桃红色精品国产亚洲av| 国产爱豆传媒在线观看| 女的被弄到高潮叫床怎么办 | 免费观看精品视频网站| 一区二区三区免费毛片| 国产 一区精品| 午夜免费男女啪啪视频观看 | 亚洲欧美日韩高清在线视频| 日韩av在线大香蕉|