• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Augmented Deep Multi-Granularity Pose-Aware Feature Fusion Network for Visible-Infrared Person Re-Identification

    2024-01-12 03:47:06ZhengShiWanruSongJunhaoShanandFengLiu
    Computers Materials&Continua 2023年12期

    Zheng Shi,Wanru Song,Junhao Shan and Feng Liu

    School of Educational Science and Technology,Nanjing University of Posts and Telecommunications,Nanjing,210013,China

    ABSTRACT Visible-infrared Cross-modality Person Re-identification (VI-ReID) is a critical technology in smart public facilities such as cities,campuses and libraries.It aims to match pedestrians in visible light and infrared images for video surveillance,which poses a challenge in exploring cross-modal shared information accurately and efficiently.Therefore,multi-granularity feature learning methods have been applied in VI-ReID to extract potential multigranularity semantic information related to pedestrian body structure attributes.However,existing research mainly uses traditional dual-stream fusion networks and overlooks the core of cross-modal learning networks,the fusion module.This paper introduces a novel network called the Augmented Deep Multi-Granularity Pose-Aware Feature Fusion Network(ADMPFF-Net),incorporating the Multi-Granularity Pose-Aware Feature Fusion(MPFF)module to generate discriminative representations.MPFF efficiently explores and learns global and local features with multi-level semantic information by inserting disentangling and duplicating blocks into the fusion module of the backbone network.ADMPFF-Net also provides a new perspective for designing multi-granularity learning networks.By incorporating the multi-granularity feature disentanglement (mGFD) and posture information segmentation (pIS) strategies,it extracts more representative features concerning body structure information.The Local Information Enhancement (LIE) module augments high-performance features in VI-ReID,and the multi-granularity joint loss supervises model training for objective feature learning.Experimental results on two public datasets show that ADMPFF-Net efficiently constructs pedestrian feature representations and enhances the accuracy of VI-ReID.

    KEYWORDS Visible-infrared;person re-identification;multi-granularity;feature learning;modality

    1 Introduction

    Person re-identification(Person ReID)aims to retrieve the same target pedestrian from various cameras with non-overlapping viewpoints.With the popularization of surveillance devices and the continuous development of computer vision,ReID has received increasing academic attention.As methods based on deep learning methods are continuously proposed,the performance of the ReID task has been highly improved[1–3].However,in the real-world application of surveillance,there are frequently low-light situations,such as at night.With the popularity of 24-h surveillance cameras,those in public places can switch to infrared mode by themselves at night.For another example,with the increasing significance of big data and artificial intelligence technology,there is a growing emphasis on providing students who will inhabit the future society with education in data science and AI[4].In the process of building a teaching environment such as a Digital Campus,the infrared camera can be added to the surveillance system to ensure the security of the place and the diversity of educational information.It seems to be a challenging issue if the retrieval of pedestrians or students is performed in the above application scenarios.Additionally,Internet of Things(IoT)based interconnected vision sensors in smart cities are considered essential elements of a cyber-physical system(CPS)and contribute significantly to urban security.However,the Re-ID of targeted persons using emerging edge AI techniques still faces certain challenges[5].Consequently,the visible-infrared cross-modality person re-identification(VI-ReID)has also emerged.

    In the visible-visible ReID (VV-ReID) task,researchers need to deal with the discrepancies between pedestrian images of the same identity(ID)to achieve accurate recognition.The discrepancies can be regarded as intra-modality differences,which are generated by changes in pose,environment,etc.The network design [6–8] and the distance metric loss [9,10] are utilized to narrow the distance of the same ID and enlarge the intra-class differences of different pedestrians.In contrast to VVReID,a significant and central challenge in VI-ReID technology lies in addressing the substantial inter-class differences caused by the modality variations between visible and infrared images.The visible and infrared images are respectively captured by reflected light and thermal radiation.Hence,there is an inherent difference between the imaging principles of both,resulting in captured images of pedestrians presenting diverse appearances in these two domains.In addition,the problems due to the short development time and the methods mostly migrated from the traditional VV-ReID are still preventing further in-depth development and research.

    In order to achieve the accuracy improvement of VI-ReID,researchers have investigated three aspects of shared feature mining [11–15]: heterogeneous distance loss function design [16–19] and heterogeneous image Translation[20–23].For feature learning,existing methods[11,16,24]generally perform coarse-grained learning from the image-level global feature.Although they can reduce the inter-modal disparity,the fine-grained representation of detailed information and body structure of pedestrians represented by local features are neglected to a larger extent.Several works [17,25,26]focused on more detailed local features through methods based on slicing,such as PCB.These methods bridge the gap between a great quantity of useful fine-grained information in cross-modal tasks.Existing approaches of local feature extraction are mostly based on average segmentation of the feature map that is outputted from the backbone.However,after extensive and detailed observation of the dataset,we also realize that it is difficult for the equal-scaled slices to describe the body structure accurately.This leads to high redundancy and low accuracy of the obtained fine-grained information.In fact,it is due to body misalignment caused by the different shooting perspectives.Moreover,it brings another problem simultaneously:excessive attention to local features may inevitably result in ignoring coarse-grained information about pedestrians,such as posture and body shape.

    Multiple Granularities Network(MGN)certainly provides a great solution to this issue[27,28].It integrates various detailed information with coarse-grained.Authors explore and improve MGN for efficient supervised learning of multiple granularity information.However,the network structure proposed in[28]is much more complicated and redundant.It is a four-stream parallel learning by two branches,namely global and local branches,so as to realize the common attention to coarse and fine granularity.In turn,this leads to the issues of huge parameters and high computational costs.The existing MGN models are all derived from ReID,while the main difference in VI-ReID is the shared feature learning process across modalities.Therefore,constructing a shared network that allows the fusion module responsible for shared learning to transmit coarse-grained and fine-grained information more efficiently and deeply becomes the key to breakthroughs in future multi-granularity cross-modal ReID research.

    To address the above issues and improve the effectiveness of the feature learning in the VI-ReID task,this paper proposes a novel framework named Augmented Deep Multi-granularity Pose-aware Feature Fusion Network(ADMPFF-Net)for mining the rich cross-modality information.ADMPFFNet is designed based on the parameter-sharing network to learn global-local features simultaneously[17].The network aims to learn multi-level shared features from pedestrians in the cross-modality task.To enable the fusion component,which serves as the core of the VI-ReID network,to learn multi-granularity information more deeply and comprehensively.The network utilizes the Multiplegranularity Pose-aware Feature Fusion(MPFF)module.The most significant difference between the Multiple-granularity Pose-aware Feature Fusion(MPFF)module and other multi-granularity feature fusion networks lies in its design approach.Traditional network designs often use a single fused module to separately extract coarse-grained and fine-grained features.In contrast,the MPFF module splits the fusion module into two parts: the first part continues to use a parameter-sharing network to fuse features,while the second part is duplicated and divided into multiple feature extraction units to achieve multi-granularity learning.This design approach enables the fusion module,which serves as the backbone of the network,to more effectively conduct deeper exploration and learn potential multi-granularity features.Additionally,this network proposes an improved slicing strategy based on the common characteristics of the body for fine-grained feature learning.This strategy chooses to use body-aware non-equal slices,thus better preserving better preserve the original body structure of humans and reducing the information redundancy caused by structural confusion.This strategy opts for non-uniform slicing based on pedestrian perception,dividing local features according to a 1:2:3 ratio for the head,upper body,and lower body.In contrast to traditional three equal parts slicing,it better preserves the original structure of the pedestrian body and reduces information redundancy caused by structural ambiguity.As a pioneering attempt in local feature extraction,its lightweight and efficient nature offers potential improvements for other methods and networks aimed at extracting local features.In order to achieve the optimization of the features that have a strong representation capability in the fine-grained information,the network is designed to replace the coarse-grained part of the global feature representation with the proposed local information enhancement(LIE)module.As a result,discrepancies among multi-granularity information can be compensated,and the effectiveness of particular feature representations can be enhanced.Thus,the accuracy of the network for multigranularity information learning can be improved.Moreover,in order to supervise the learning process of multi-granularity features,we design a multi-granularity joint(MGJ)loss.The model’s loss is based on both traditional classification loss and triplet loss,enhancing its robustness.

    In a nutshell,the contributions of this paper are summarized as follows:

    ? This paper proposes ADMPFF-Net for simultaneously learning multi-level features for VIReID.By deeply integrating multiple granularity units into a shared feature module,the model learns more robust representations in the cross-modality situation under the supervision of the MGJ loss.

    ? MPFF provides a novel and efficient backbone type for multi-granularity feature learning methods.Furthermore,combined with the innovative non-equal slicing strategy,it not only achieves precise and deep mining of multi-granularity features but also offers a promising avenue for related networks focused on local feature learning.

    ? LIE is designed to replace the global feature independent learning branch.It augments the learning of stable local features and solves the issue of model oscillations caused by different information scales.

    ?Extensive experiments are conducted on the two public datasets,and the results validate that the proposed method can achieve state-of-the-art performance on both datasets.

    The rest of this paper is organized as follows.Section 2 discusses the work of VI-ReID related to this paper,and Section 3 gives an introduction to ADMPFF-Net in detail.To demonstrate the effectiveness of the proposed network,we present numerous experiments in Section 4,including both comparison experiments and ablation experiments.And finally,we bring a conclusion to the method in Section 5.

    2 Related Work

    This section provides an overview of VI-ReID methods.The fundamental goal of all these methods is to minimize cross-modal discrepancies to the greatest extent possible while enhancing intra-class differences.Currently,the studies in the field of VI-ReID can be roughly divided into feature learning,metric learning,and image translation[2,3].In particular,multi-granularity shared feature mining has been valued in the VI-ReID task as a new research direction in feature representation.

    2.1 Feature Learning

    The essence of feature learning is to obtain more effective and robust representations of pedestrians in the cross-modal task.This approach primarily focuses on designing efficient network structures to learn and extract shared feature representations that are most suitable for two modalities of images,aiming to reduce the differences between modalities.Wu et al.[29] defined the cross-modal person ReID in 2017 and divided models into single-stream,dual-stream,and asymmetric fully connected layers.Ye et al.[18,30] employed the dual-stream network to separately extract shared information from the two modalities.They achieved feature fusion through feature mapping to the same space during the feature embedding process.Building on this foundation,Liu et al.[13]merged intermediate layer features to enhance cross-modal information integration.Zhang et al.[31]introduced a dual-path modality framework for feature extraction.This approach preserved the dual-path spatial network and its structure in the common space while focusing on the co-variance of input cross-modal images to deeply explore feature correlations.Additionally,the methods mentioned above,attention mechanisms have also been applied to assist in the process of feature learning.For example,Wang et al.[32]proposed an Attention-based Dual Stream Modality-aware method(ADSM)to effectively learn the same-modality inter-class differences.However,although feature representation learning has made certain progress,its performance on large-scale datasets has not yet reached a high level.

    2.2 Metric Learning

    Different from feature learning,metric learning not only mines shared information through various network frameworks,but also requires the design of an effective metric learning method or distance metric loss function.These methods aim to constrain the training process of the model and narrow the distances for single-and cross-modal cases.Ultimately,more effective cross-modal ReID models are formed for accurate detection and identification.Wu et al.[24] proposed a focusaware similarity preservation loss to guide the learning of modality-similar features.Zhao et al.[15]introduced a Hard Quintuplet Loss,composed of global triplet loss and difficult cross-modal triplet functions.Similarly,Liu et al.[13] designed a Dual-Modal Triplet Loss.All of these approaches contributed to improving the accuracy of VI-ReID to a certain extent.

    The concept of center loss has been widely employed in VI-ReID.For instance,Zhu et al.[16]introduced the Asymmetric Center Loss,which reduced cross-modal differences between classes and avoided excessive computation.Additionally,more recent improvements were based on heterogeneous centers and difficult triplet losses.Liu et al.[17]proposed the Heterogeneous Center Triplet Loss and similar approaches.These methods optimized learning through the design of loss functions or network models.However,most of the above methods employed simple slices and weightless constraints on local features,and there remains limited focus on addressing the loss due to multi-granularity information differences.

    2.3 Image Translation

    Kniaz et al.introduce a new direction in VI-ReID through image generation based on Generative Adversarial Networks (GANs) [33].They transformed visible light images into infrared images to supplement the dataset.In [34],the authors utilized the adversarial training concept of GANs,proposing a cross-generative adversarial network based on adversarial training to eliminate variations in cross-modal images.To bridge the cross-modal gap,in[35],Wang et al.presented a relief learning framework based on bidirectional cycle generators.Choi et al.[36]proposed the Hierarchical Intra-Cross-Modal Disentanglement(Hi-CMD)method,which automatically extracted key identification information from both infrared and visible modalities to reduce the impact of cross-modal information disparities.Wei et al.[37] proposed the Flexible Body Partition (FBP) model-based adversarial learning method(FBP-AL)to focus on designing modality classifiers to automatically distinguish part representations for obtaining more effective modality-sharable features.This also included the recent work by Zhong et al.[23],who proposed the Grayscale Enhancement Colorization Network(GECNet)for VI-ReID.It assisted in learning cross-modal image information by generating intermediate grayscale images.

    However,although the above methods have improved recognition efficiency and accuracy,imagebased generation methods still suffer from uncertainties in performance due to the disregard of critical details(such as color information)and limitations in model access.

    2.4 Multi-Granularity Shared Feature Mining

    The current methods of feature learning ignore the enormous potential inherent in multigranularity information.Significant achievements have been made in the field of VI-ReID to realize multi-granularity feature learning by simultaneously focusing on local and global features.As illustrated in the GLMC network by [28],it has been attempted to extract multi-granularity features to reconcile global and local parts by focusing on both coarse and detailed information simultaneously.The study by Tang et al.[8]summarized the general approach to local feature extraction,which was to partition the output feature of the network on a horizontal scale.This research suggested that the method is too coarse.It not only overemphasized the noise information but also destroyed the scale structure of the normal pedestrian.

    While existing methods have used the MGN network from person VV-ReID to achieve multigranularity feature learning,they have not considered the key issue in VI-ReID,which is addressing the substantial inter-class differences.Our designed network takes this into account and utilizes an innovative disentanglement approach to enable the fusion module to learn multi-granularity features more efficiently and accurately.It not only improves the application of multi-granularity feature learning in VI-ReID but also provides a new direction for designing feature networks with multigranularity and multi-scale learning capabilities.

    3 Proposed Method

    The architecture designed for VI-ReID,termed Augmented Deep Multi-granularity Pose-aware Feature Fusion Network (ADMPFF-Net),is illustrated in Fig.1.Inspired by HcTri [17],the main architectural framework of this network is based on the classical dual-stream architecture.Therefore,the method proposed in HcTri [17] can be regarded as the baseline of VI-ReID.The backbone of ADMPFF-Net is ResNet-50,which is pre-trained on the ImageNet dataset.ADMPFF-Net consists of four main modules,namely(1)dual-stream specific and shared feature extraction module,(2)multigranularity pose-aware feature fusion module,(3) local information enhancement module,and (4)multi-granularity joint loss module.The integration of these modules aims to achieve discriminative and adaptive multi-granularity feature representation learning in the VI-ReID task,thus enhancing the accuracy of VI-ReID.The following subsections provide detailed descriptions of these key techniques.

    Figure 1:Framework of the proposed ADMPFF-Net,which includes the feature separation module represented by light blue,feature fusion represented by yellow,the disentanglement part represented by green and dark blue,and the augmented part represented by orange.Rectangles of the same color represent feature stripes within the same branch.The arrows point to the losses applied to the feature stripes at the end

    3.1 Dual-Stream Network for Specific and Shared Feature Learning

    The dual-stream feature extraction network is widely used to extract cross-modal information[30].It can be divided into two parts:feature extraction and fusion.The feature extractor is responsible for obtaining shallow-level specific features from the two heterogeneous inputs by using two separate branches.The feature fusion part then projects these specific features into a common space to uncover more effective modal-shared features.However,existing research just employs feature fusion through sharing the last convolution layers or fully connected layers,thus resulting in shallow and less efficient feature fusion.

    To address the above-mentioned issue,during the specific design process of the network,stages 0 toiof the model serve as feature extraction layers to process the input images.Therefore,the shallow modality-specific features from the two different modalities can be obtained.The visible light and infrared feature extraction branches are represented as functionsfuiandfvi,respectively.The dualbranch network takes visible and infrared images as input informationxandy,resulting infui(x)andfvi(y).Here,irepresents the stage from which shared parameter layers start.Wheni=1,it indicates the layers beforestage1serve as the feature extractor.After obtaining modality-specific features,feature embedding is necessary to achieve feature fusion and obtain highly discriminative modality-shared features.A common approach is to concatenate the two feature vectors as input to the feature fusion module.To projectfui(x) andfvi(y) to the same subspace,forming the input feature for the fusion module,a unique spatial computation is required.This ultimately results in the final fused featureg(s),as expressed below:

    In Eq.(1),the parametersiandjjointly determine the specific stages of the fusion module.Here,irepresents the starting stage of the fusion module,whilejrepresents the ending stage.For example,wheni=1 andj=1,it indicates that stage 1 serves as the fusion layer.The functiong(·)denotes the shared feature learning space for cross-modal features,which is essentially feature fusion.The variable s represents the cross-modality fused feature formed by connecting the two modality-specific features from the dual-stream extractor.Finally,the obtainedg(s)is fed into the subsequent modules of the proposed ADMPFF-Net for deep feature extraction.

    3.2 Multiple-Granularity Pose-Aware Feature Fusion

    When both coarse-and fine-grained information is deeply integrated into the model,there is a problem of model oscillation that tends to be magnified.This is probably due to the large differences between the information content of the different granularity features.Moreover,existing research typically involves the horizontal proportional segmentation of network output features.Nevertheless,this method may overly focus on the noisy information,thereby destroying the body structure in normal proportions of the pedestrian.Therefore,ADMPFF-Net introduces Multiple-granularity Pose-aware Feature Fusion (MPFF) module for VI-ReID.Instead of employing the global feature,MPFF adopts a strategy that connects segmented local features to form a complete coarse-grained output.It consists of two main components,as shown in Fig.1 below,namely the multi-granularity feature disentanglement strategy(mGFD)and the posture information segmentation strategy(pIS).These strategies correspond to two multi-granularity feature extraction branches and one local feature enhancement branch.Each branch not only strips local features from the image but also connects all of them to form a global representation.The local feature enhancement branch segments the feature using the corresponding method and selects the high-performance representation as the augmented part of the local feature.For all feature strips,a 1×1 conv block is used to reduce the dimensionality of the features.

    3.2.1 Multi-Granularity Feature Disentanglement Strategy

    We explainmGFD for Strategy 1 shown in Table 1.In table,theDisentanglement part of the first stagedenotes the stage after this layer as the unshared parameter module copied for extracting multigranularity information;stage0–1denotes the layer ofstage0tostage1as the stage contained in this module.For scheme 1,we set thestage0as a dual-stream specific feature extraction module for crossmodal images;stages1–3are utilized to obtain the shared feature with feature fusion,and the finalstage4is copied into three independent subspaces for disentangling and learning different granularity information.Two of these subspaces serve as multi-granularity modules,while one functions as the local feature enhancement module.Schemes 2,3,and 4 follow a similar pattern,each with its respective purpose.

    Table 1:Different combinations for mGFD

    The introduction ofmGFD reduces the disruption of structural information by different detailed data.We denote the multi-granularity feature module ash(·).Thus,local information processing is represented byhl1(xl1i) andhl2(xl2i),wherexl1iandxl2idenote thei-thlocal features,which come from the three-part slice branch and six-part slice branch,respectively.The process of enhancing the local feature can be denoted ash(xen),wherexencorresponds to the selected enhanced local feature in experiments.Consequently,the output feature map of the proposed network is represented by:

    The details about the subsequent processing of the output features by the network are illustrated in Fig.2.It is mainly divided into 3 major steps.(1)The max-pooling layer is employed to reduce the dimensionality of the original features;(2)The 1×1 conv block is utilized to reduce the feature channel dimension(dim);(3)The batch Normalization layer and ReLu layer are used to make the training more generalized and stable.In the end,the construction of the complete test feature representation is as shown in the following equation:

    In order to better construct a test feature that comprehensively reflects the overall network learning status and at the same time ensures that the constructed feature is more stable and conforms to the correct human structure,the construction process begins by concatenating the complete pedestrian featuresandseparately from the final outputs of the two local branches.Then,the enhanced feature is inserted at appropriate positions.As shown in the equation above,the leg part h(xen)is used as the enhancing feature and is inserted at the bottom of the complete pedestrian feature.

    Figure 2:Implementation details of multi-granularity feature disentanglement strategy

    It can be observed thatmGFD is different from conventional strategies for global feature learning.Existing strategies are developed from VV-ReID and utilize independent branches for global feature extraction.ThemGFD module selectively segments local features into sub-features,which are then fused to generate global features.Specifically,it employs two different branches to extract three and six sub-features,respectively.After subdivision,these features are recombined to form a complete global feature,each of which is subject to an independent loss constraint.Additional details about these features can be seen in Table 2.This strategy on the one hand enables comprehensive multi-granularity information extraction,and,on the other hand,enhances the stability of the model parameters.

    Table 2:The number of feature strips,the feature map size and the corresponding dimensions after dimensionality reduction by a 1 ?1 convolution block are output by the three independent branches of the disentanglement module

    3.2.2 Posture Information Segmentation Strategy

    The most common method for extracting local features is to divide the feature map into equal parts.The parameterpcan be set to different values,indicating that the pedestrian features are divided into various numbers of horizontal parts.For example,whenp=3,the three local features correspond to the head,upper body,and lower body of the pedestrian image,respectively.And whenp=6,the aim is to provide a more detailed description of the body,such as the chest or abdomen.The results of the above-mentioned segmentation are depicted in Fig.3 below.It is evident that the segmentation method whenp=3 encounters issues with unreasonable splitting,leading to inaccurate representation of specific body parts in the local features.As illustrated by the red and blue boxes in Fig.3,segmentation that ignores pose information disrupts the inherent body structure of the person.Misalignment in segmentation introduces interference between different body parts.The introduction of redundant information impacts training efficiency and further reduces the accuracy of the model.

    Figure 3:The difference between pose-aware segmentation and equal segmentation is demonstrated in(a)and(b).Pedestrians in(a)and(b)are from visible and infrared modalities,respectively.The red and blue boxes represent the torso and lower body of the original human structure,respectively

    By analyzing the pedestrian images in the VI-ReID dataset,it can be observed that whenp=3,the typical proportions of a pedestrian’s head,upper body,and lower body in the images are approximately 1:2:3.Based on it,we adopt a novel feature segmentation strategy namedpIS for local feature extraction.Its main purpose is to partition the output features into different stripes in the ratio of 1:2:3,which corresponds to the normal proportions of the human head,torso,and lower body.Compared with other non-equal segmentation strategies,such as the method proposed in [8],which combines fused features and gait energy images(GEI),pIS not only accurately captures complex body details,but also avoids introducing excessive parameters.Consequently,it can be applied more effortlessly and efficiently in multi-scale feature learning tasks.

    3.3 Local Information Enhancement Module

    In order to emphasize the role of important information,we designed a LIE module by adding a detached branch for VI-ReID.Obviously,there is a great difference in the distinctiveness and saliency of local features at different locations.Therefore,to emphasize the highly expressive local feature,we introduced a LIE module for the cross-modality task.In the detail of the model design,an independent branch for local feature enhancement is proposed following the process of themGFD strategy.This enables repetitive reinforcement learning of the specific feature.

    Based on observations from datasets and extensive exploration of local features,we analyze the representational ability of local parts of pedestrians,and the results are presented in Fig.4.It uses the green boxes to highlight regions with strong representational capabilities,while the red boxes represent structures that are less discriminative for different pedestrians.When the local features are divided into three parts,the features of the head and lower body are represented significantly higher than the torso.Similarly,the local features perform relatively well when divided into six parts,except for the region of the thighs and abdomen.Therefore,it is important to reinforce the critical region from the shallow layer of the network using a separate branch to improve the model discrimination.we analyze the representational ability of local parts of pedestrians,and the results are presented in Fig.4.

    Figure 4:The red boxes are labeled as lower-resolution structures and the green boxes are labeled as higher-resolution structures

    The attention mechanism adjusts the weights by automatically calculating the feature weights through training.However,it tends to determine the importance of deep-level features.Therefore,attention mechanism may ignore the important role played by mid-level semantic information in judging whether it is the same pedestrian or not.Unlike the attention mechanism,LIE is based on the analysis of pedestrian images and strengthens from the mid-level features,which is more consistent with the description of human structure.In the paper’s experiments,we show the performance difference between using the attention mechanism and the LIE module.

    3.4 Multi-Granularity Joint Loss

    In order to further enhance the similarity between the same pedestrian,a novel metric loss named multi-granularity joint(MGJ)loss is proposed for VI-ReID.We use ID loss and heterogeneous center triplet loss (Lhc_tri) to optimize the learning of the model.The ID loss is mainly utilized to supervise the identity information predicted by the fully connected (FC) layer and improve the accuracy and effectiveness of the classification.The detailed calculation formula is as follows:

    whereyrepresents the correct identity label,serves as the representation of the correct probability,andNis the number of IDs in the entire dataset.Additionally,ξ is a constant,which is introduced to improve the domain adaptability of the network model and is generally set to 0.1 in experiments.

    However,it is difficult to just rely on the ID loss to enhance the discrimination of the network in handling homogeneous and heterogeneous information simultaneously.For this reason,the triplet loss is widely used in the Re-ID task,which was originally introduced in FaceNet[38].It is subsequently improved through various methods,such as the mining hard triplets [39].Lhc_triused in this paper significantly differs from the above-mentioned losses.The significant advantage of it is the introduction of intra-and inter-modality centers to constrain the feature distribution.The representation of feature centers for each modality can be denoted as follows:

    Under the single granularity condition,these losses jointly lead to good performance.However,during multi-granularity feature learning,the information content of coarse-and fine-grained features is different due to varying slice sizes.In turn,this may cause the network to fall into oscillations.In order to enhance the stability of learning and enable the model to converge faster,we introduce the multi-granularity joint loss for VI-ReID.The underlying principle of the proposed loss is also similar to the loss used in mainstream multi-granularity methods.By jointly constraining the fine-grained local and coarse-grained global features with ID loss and triplet loss,respectively.The details are as follows:

    whereLmGFD_3andLmGFD_6represent the losses used in the two main branches of the mGFD module,respectively.Σ(Ltri3_i+Lid3_i) represents a simple summation of the ID loss and triplet loss of the three local feature strips.Lid3_allandLtri3_allare used to individually constrain the local feature strips that are formed by the joining of the global features.In fact,for each of the six local feature stripes,we utilize the above method to calculate the loss.The above loss methods only achieve simple constraints on multi-granularity features.The highlight of the MGJ Loss proposed in this paper is to balance the difference of multi-granularity information by proportioning the weight parameters.The specific implementation is as follows:

    The two weight parameters α and β are mainly used to balance the difference between coarse-and fine-grained information.TheLid_enandLtri_enrepresent the loss functions used by the LIE Module.The MGJ loss enhances the stability of the model by balancing the learning process of multi-granularity features.

    4 Experiments

    By conducting comparison and ablation experiments,this section mainly evaluates the performance improvements of the proposed method on two publicly available datasets,including RegDB and SYSU-MM01.It also describes the detailed experimental settings and experimental criteria.

    4.1 Experimental Settings

    4.1.1 Datasets

    SYSU-MM01serves as a large-scale cross-modal pedestrian dataset.It comprises data from 6 cameras,including 4 visible cameras and 2 infrared cameras.The training set contains 395 pedestrians,including 22,258 visible light images and 11,909 infrared images.And the testing set consists of an additional 96 individuals,including 3,803 infrared images for queries and 301 randomly selected visible images for the gallery set.Additionally,this dataset is positioned in indoor and outdoor settings.These two search modes represent distinct challenges.The evaluation involves 10 gallery set selection trials and reports the average retrieval performance.Detailed descriptions of the evaluation protocol can be found in[17].

    RegDBis another extensively used dataset.It is established by a dual-camera system(one visible camera and one thermal camera) and contains 412 individual identities.Each person has 10 visible images and 10 thermal images.Following the evaluation protocols used in[17],the dataset is randomly divided into training and testing subsets.During testing,images from one modality(thermal images)are utilized as the gallery set,while images from the other modality(visible images)serve as the query set.This process is repeated 10 times for result stability.

    4.1.2 Evaluation Protocol

    During the training process,we adopt Cumulative Matching Characteristics (CMC),mean Average Precision (mAP),and mean Inverse Negative Penalty (mINP) [2] as evaluation protocols for the effectiveness of the model.It is worth noting that before testing,the features of all IDs are first subjected to L2 Norm to ensure the consistency and comparability of the data.These evaluation protocols validate the mode’s effectiveness and provide powerful support and reference for this study.

    4.1.3 Implementation Details

    The proposed network is implemented on the NVIDIA RTX4090 GPU with PyTorch.Based on previous research and experimentation,the network selected a pretrained ResNet50 with ImageNet weights as the backbone network.In order to capture more detailed information,we adjusted the step size of the final convolution block during training from 2 to 1.This adjustment enlarges the feature map by 4 times.During the training process,the size of the input image is set to 288×144 with a zero padding of 10 pixels around the image.To enhance training data diversity,random horizontal flipping is employed for processing the inputting,followed by cropping to the designated size.Such data augmentation strategy contributes to improving model generalization and stability.

    For optimizer selection,we employ the stochastic gradient descent(SGD)optimizer.The momentum is set to 0.9,and the initial learning rate is set to 0.1.Furthermore,the experiment utilizes a warmup learning rate strategy,which helps to speed up the convergence of the network and achieve superior performance.The specific learning rate settings are presented below:

    Other hyper-parameters are outlined in Table 3.Moreover,the total number of training epochs is 80.

    Table 3:Hyper-parameters of the network

    4.2 Ablation Experiments

    In order to demonstrate the effectiveness of the method proposed in this paper,we perform ablation experiments on public datasets.The experiments are employed to evaluate the effectiveness of different components in ADMPFF-Net,including MPFF,LIE,and multi-granularity joint loss.

    4.2.1 Effectiveness of MPFF Module

    MPFF consists of two main components: Multi-granularity Feature Disentanglement Strategy(mGFD)and Posture Information Segmentation Strategy(pIS).In the following,specific experiments verify the effectiveness of the two strategies in multi-granularity feature learning,respectively.

    Effectiveness of mGFD Strategy:The central part of themGFD strategy is the design of the disentanglement module.The key to effectively extracting deep multi-granularity features lies in the appropriate stage selection of the disentanglement module.As shown in Table 4,when the shared module is configured asstages2–4without the disentanglement module,the performance on SYSUMM01 is significantly lower than the others.This indicates thatmGFD effectively promotes feature learning.Furthermore,the experimental results show that when the fusion modules become too large,the accuracy of the model in handling multi-granularity features may decrease,leading to a decline in performance.This aligns with the results where the best performance is achieved when stages 1–2 are set as the shared module(stages 3–4 as the disentanglement module).Compared to other settings of the disentanglement module,SYSU-MM01 shows a significant improvement of 3% to 4%.This finding emphasizes the importance of reasonable control over the disentanglement module in multigranularity feature fusion and provides further experimental validation for the approach proposed in this paper.

    Table 4:Experimental results for different fusion part under the rank 1,mAP,and mINP(%)criteria for the SYSU-MM01 datasets.Stages denote the layers of the fusion part

    Effectiveness of pIS Strategy:ThepIS strategy offers a new approach for extracting more discriminative local features.When compared to traditional equal segmentation methods,even finer divisions such as 6 parts or 9 parts,pIS proposed in this study can explore human body structural information and perform feature segmentation based on more precise and reasonable proportions.As depicted in Table 5,pIS can be observed as an improvement of nearly 3% in rank-1 on the SYSUMM01 dataset.Additionally,achieves even more significant enhancements in the mAP and mINP accuracy,with improvements exceeding 9%.Hence,in comparison to mainstream strategies,thepIS segmentation strategy exhibits significant advantages in terms of computational cost and feature segmentation efficiency.

    Table 5:Results of the first branch under various segmentation strategies on the SYSU-MM01 dataset

    4.2.2 Effectiveness of LIE Module

    To explore the effectiveness and optimal performance of the LIE Module,the following experiments were conducted,focusing on both network branch configuration and module feature selection.

    Effectiveness of Network Branch Configuration:ADMPFF-Net focuses on emphasizing features with better representation.Mainstream networks often employ separate global branch(Global part))to fully extract coarse-grained information.As shown in Fig.5 below,such methods achieve results of 63.32% for rank-1,61.47% for mAP,and 48.04% for mINP accuracy.In contrast,this paper proposes using the LIE branch as an alternative.On the one hand,the attention mechanism attempts to focus on features with higher expressiveness through the learning of feature weights but ignores the mid-level semantic information.On the other hand,our proposed LIE module starts from the mid-level features and enhances the attention to the more structurally superior features.To verify the effectiveness of the LIE branch,the study conducts a series of experiments to compare with different cases.The cases use the attention mechanism branch (Attention),the global feature branch (Global part) and non-branch (Non).As shown in the results in Fig.5 below,the performance of the LIE branch configuration significantly surpasses other network branch configuration methods on rank-1,mAP and mINP accuracy.

    Figure 5:Results on SYSU-MM01,the cases with the Attention,Global part and Non are employed to compare with LIE

    Effectiveness of Feature Selection in LIE:The feature selection experiments for the LIE module involve two major steps: identifying specific high-performance structures and determining the corresponding feature sizes.Firstly,as indicated in Fig.6a,the feature of the torso has relatively low discrimination.In comparison,the head and leg features exhibit better performance and reflect human body information more effectively.Secondly,by applying different segmentation sizes to the optimally performing leg features,the optimal feature map learning size is determined.Combining the experimental results in Fig.6b,it is observed that on the SYSU-MM01 dataset,the leg feature size of 3×9 yields the optimal performance across various indicators.

    Figure 6:(a)Depicts the effectiveness of different structural features extracted from the human head,upper body,and lower body;(b)presents performance variations under different sizes that explore the optimal structure

    4.2.3 Effectiveness of the MGJ Loss

    In this subsection,related experiments primarily validate the effectiveness of our proposed Multigranularity Joint Loss from two perspectives.On the one hand,different from other multi-granularity feature learning methods,the coarse-grained part of ADMPFF-Net is generated through recombining and connecting local features.It is individually constrained by a specialized loss.Therefore,it is essential to demonstrate the efficacy of this individual loss in constraining coarse-grained information.On the other hand,it is important to establish the superiority of the proposedLMGJloss over traditional triplet lossLtriand hetero-center triplet lossLtri_hcin constraining multi-granularity features by enhancing inter-class separability and intra-class compactness.

    Effectiveness of Coarse-Grained Loss:This experiment evaluates the efficiency of multi-granularity feature extraction compared to single local or global features under the same network architecture and parameter settings.In the specific loss construction process,we primarily utilize individual losses to constrain the global features formed by connected local features,achieving the Coarse-grained Loss.The result in Table 6 presents that just using local features outperforms employing global features due to the effective enhancement from LIE.Moreover,the most optimal results are achieved by individually constraining the coarse and fine-grained losses,showcasing improvements of over 4%on rank-1,mAP and mINP compared to others.This robustly demonstrates the superiority of learning multi-granularity information.

    Table 6:“yes” indicates that we have configured the features at that granularity level,while “no”signifies the opposite.We have analyzed the specific results under various granularity loss settings

    Effectiveness of LMGJ:In this experiment,a comprehensive comparison is conducted between the constraint effects of triplet lossLtriand hetero-center triplet lossLtri_hc,inI am running a few minutes late;my previous meeting is running over.contrast to our proposed Multi-granularity Joint LossLMGJ.As depicted in Table 7,LMGJexhibits significant performance enhancement compared to the unweightedLtri_hc.Moreover,Ltrieven fails to converge properly on the SYSU-MM01 dataset.This not only verifies the superiority of the selected hetero-center triplet lossLtri_hc,but also emphasizes that the awellocation of weighted parameters effectively mitigates learning confusion arising from varying feature scales.This underscores the superiority ofLMGJin handling multi-granularity information tasks.

    Table 7:Results on SYSU-MM01 are constrained by different loss functions

    4.3 Comparison with State-of-the-Art Methods

    This section primarily focuses on a comprehensive comparison between our proposed ADMPFFNet and several SOTA VI-ReID methods.A detailed comparative analysis on the RegDB and SYSUMM01 datasets is presented in Tables 8 and 9.The experimental results demonstrate that the proposed ADMPFF-Net can significantly improve the recognition and learning of cross-modal pedestrian images.

    The experiments on the RegDB dataset are shown in Table 8,where the proposed method achieves the best performance in both query modes.The performance inVisible to Thermalmode is even optimized to rank1/mAP/mINP with 93.54%/85.02%/70.82%,which is an effective improvement over other methods.For example,compared to GECNet[23],ADMPFF-Net demonstrates performance advantages of 11.21% in rank-1 and 6.57% in mAP.When compared to the baselineHcTri[17],ADMPFF-Net exhibits performance improvements of 2.49% in rank-1,1.74% in mAP,and 1.98%in mINP.This simultaneously demonstrates the effectiveness of the multi-granularity loss used in this paper in guiding multi-granularity learning.Even when compared to the latest methods such as FBP-Al[38],the improvements in various metrics are also quite significant.These studies validate the effectiveness of the multi-granularity feature learning framework used in this research.

    Table 8:State-of-the-art methods on the RegDB dataset were compared,rank-1,mAP,and mINP(%)

    On the SYSU-MM01 dataset,we compare various methods with ADMPFF-Net,and the specific results are presented in Table 9.From the table,it can be observed that the proposed method achieves notable improvements over the baseline method HcTri [17].The rank-1,mAP,and mINP accuracy can be improved from 61.68%,57.51%,and 39.54% to 67.66%,63.5%,and 49.70%,respectively.Additionally,in comparison with GLMC [28],a method with a larger parameter size but a similar focus on multi-scale information extraction,our proposed method achieved a 3.29% improvement in the rank-1 accuracy.Furthermore,when compared to the latest methods such as ADSM [34],ADMPFF-Net exhibits significant superiority,with improvements in relevant metrics exceeding 10%.What’s more,it substantially reduces the relative computational load and computation time.Therefore,ADMPFF-Net making it exceptionally outstanding for practical applications.

    Table 9:State-of-the-art methods on the SYSU-MM01 dataset were compared,rank-1,mAP,and mINP(%)

    5 Conclusion

    This study introduces ADMPFF-Net,a novel network designed to optimize the VI-ReID model for extracting discriminative and robust multi-granularity information about pedestrians.ADMPFFNet improves upon the dual-stream feature learning network by introducing the MPFF module.This module incorporates the mGFD andpIS strategies,allowing the extraction of both coarse-and finegrained information from pedestrian images in a cross-modal setting.These strategies greatly enhance the efficiency and accuracy of mining multi-granularity information,offering an innovative and effective approach to designing multi-granularity feature learning networks.In addition,ADMPFFNet can be applied as a novel baseline for recognition tasks in scenarios of street,station and campus.

    Furthermore,the proposed method integrates the LIE module,which adds separate branches behind the middle layer of the network.It enhances the role of high-performance features and promotes stability in the overall feature-learning process.

    To facilitate effective feature learning in VI-ReID,the proposed multi-granularity joint loss is employed to supervise model training.This loss function focuses on extracting more discriminative multi-granularity features.Through extensive experiments conducted on two prominent VI-ReID datasets,the results demonstrate that the proposed method effectively constructs pedestrian feature representations,leading to improved recognition accuracy.

    Acknowledgement:We want to thank the team at Sun Yat-sen University for creating the SYSU dataset and the Hong Kong Polytechnic University for providing the RegDB dataset.This has laid a critical foundation for our research and experiments.

    Funding Statement:This work is supported in part by the National Natural Science Foundation of China under Grant 62177029 and 62307025,in part by the Startup Foundation for Introducing Talent of Nanjing University of Posts and Communications under Grant NY221041,and in part by the General Project of The Natural Science Foundation of Jiangsu Higher Education Institution of China 22KJB520025 and 23KJD580.

    Author Contributions:The authors confirm contribution to the paper as follows:study conception and design:Z.Shi and W.Song;data collection:F.Liu and J.Shan;analysis and interpretation of results:Z.Shi and W.Song;draft manuscript preparation:W.Song,Z.Shi.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The additional datasets generated during and analyzed during the current study are available from the corresponding author upon reasonable request.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    色噜噜av男人的天堂激情| 亚洲经典国产精华液单| 色综合站精品国产| 成人一区二区视频在线观看| 欧美xxxx性猛交bbbb| 1024手机看黄色片| 国产真实乱freesex| 少妇被粗大猛烈的视频| 卡戴珊不雅视频在线播放| 丝袜美腿在线中文| 欧美在线一区亚洲| 乱系列少妇在线播放| 免费黄网站久久成人精品| 久久精品国产自在天天线| 三级男女做爰猛烈吃奶摸视频| 国内精品美女久久久久久| 精华霜和精华液先用哪个| eeuss影院久久| 久久精品国产清高在天天线| av专区在线播放| 成人综合一区亚洲| 不卡视频在线观看欧美| 老熟妇乱子伦视频在线观看| 免费在线观看影片大全网站| 人人妻人人澡欧美一区二区| 一区福利在线观看| 亚洲18禁久久av| 亚洲av第一区精品v没综合| 午夜精品在线福利| 麻豆av噜噜一区二区三区| av在线老鸭窝| 伦精品一区二区三区| 国产精品国产高清国产av| 99热这里只有是精品50| 精品一区二区三区av网在线观看| 在线国产一区二区在线| 网址你懂的国产日韩在线| 亚洲中文字幕一区二区三区有码在线看| 国产精品久久久久久亚洲av鲁大| 国产一区二区在线观看日韩| 99热只有精品国产| 国产黄色小视频在线观看| 搡老岳熟女国产| 黄色欧美视频在线观看| 欧美激情久久久久久爽电影| 性插视频无遮挡在线免费观看| 亚洲av中文av极速乱| 97超碰精品成人国产| 久久久精品大字幕| 成人性生交大片免费视频hd| 国产精品精品国产色婷婷| 国产精品一及| 一区福利在线观看| 丰满人妻一区二区三区视频av| 久久精品综合一区二区三区| 国产亚洲av嫩草精品影院| 成熟少妇高潮喷水视频| 国产成人a区在线观看| 一边摸一边抽搐一进一小说| 国产精品av视频在线免费观看| 色5月婷婷丁香| 久久天躁狠狠躁夜夜2o2o| 亚洲经典国产精华液单| 国产精品久久久久久久久免| 我要看日韩黄色一级片| 欧美精品国产亚洲| 日本 av在线| av中文乱码字幕在线| 97碰自拍视频| 精品一区二区三区视频在线观看免费| 又黄又爽又免费观看的视频| 两性午夜刺激爽爽歪歪视频在线观看| 欧美高清成人免费视频www| 哪里可以看免费的av片| 日本免费a在线| 免费观看在线日韩| 国产白丝娇喘喷水9色精品| 日韩在线高清观看一区二区三区| 国产精品野战在线观看| 在线播放国产精品三级| 内地一区二区视频在线| 精品久久久久久成人av| 精品日产1卡2卡| 晚上一个人看的免费电影| 男人狂女人下面高潮的视频| 国产v大片淫在线免费观看| 国产精品久久久久久久久免| 男人狂女人下面高潮的视频| 免费看光身美女| 日本五十路高清| 久久久久久国产a免费观看| 精品免费久久久久久久清纯| 日韩精品有码人妻一区| 亚洲国产色片| 国产高清有码在线观看视频| 丰满人妻一区二区三区视频av| 亚洲国产精品sss在线观看| 国产高清视频在线播放一区| 日日摸夜夜添夜夜添小说| 国产69精品久久久久777片| 偷拍熟女少妇极品色| 午夜a级毛片| 欧美+日韩+精品| 午夜免费激情av| 国产午夜精品久久久久久一区二区三区 | 在线观看免费视频日本深夜| 一级黄色大片毛片| 欧美不卡视频在线免费观看| 可以在线观看毛片的网站| 99热6这里只有精品| 美女黄网站色视频| 国产三级中文精品| 伊人久久精品亚洲午夜| 亚洲国产精品久久男人天堂| 亚洲美女搞黄在线观看 | 久久精品国产自在天天线| 看非洲黑人一级黄片| 一区福利在线观看| 久久久成人免费电影| 亚洲18禁久久av| 久久久久久久午夜电影| 乱人视频在线观看| 91久久精品国产一区二区成人| 夜夜爽天天搞| 久久午夜福利片| 久久精品国产99精品国产亚洲性色| 色5月婷婷丁香| 深夜a级毛片| 日日摸夜夜添夜夜爱| 蜜桃久久精品国产亚洲av| 亚洲经典国产精华液单| 色哟哟·www| 少妇人妻精品综合一区二区 | 亚洲av中文字字幕乱码综合| 精品人妻视频免费看| 亚洲av第一区精品v没综合| 欧美中文日本在线观看视频| 网址你懂的国产日韩在线| 精品国产三级普通话版| 久久久色成人| 国产伦一二天堂av在线观看| 菩萨蛮人人尽说江南好唐韦庄 | 男女做爰动态图高潮gif福利片| 嫩草影院新地址| 国产单亲对白刺激| 国语自产精品视频在线第100页| 久久久久久伊人网av| 欧美xxxx黑人xx丫x性爽| 久久久a久久爽久久v久久| 永久网站在线| 综合色av麻豆| 亚洲av电影不卡..在线观看| 成人特级黄色片久久久久久久| 国产亚洲91精品色在线| 极品教师在线视频| av.在线天堂| 五月玫瑰六月丁香| 最新在线观看一区二区三区| 三级经典国产精品| 久久人人爽人人爽人人片va| 国产美女午夜福利| 国产中年淑女户外野战色| 国产亚洲精品久久久com| 中国国产av一级| 久久久国产成人精品二区| 91精品国产九色| 国内久久婷婷六月综合欲色啪| 亚洲av五月六月丁香网| avwww免费| 亚洲aⅴ乱码一区二区在线播放| 99在线视频只有这里精品首页| 精品99又大又爽又粗少妇毛片| 国产精品亚洲美女久久久| 男人舔奶头视频| 久久精品影院6| 熟女人妻精品中文字幕| 中国美白少妇内射xxxbb| 国产三级中文精品| 免费看美女性在线毛片视频| 美女高潮的动态| 日韩一区二区视频免费看| 久久久午夜欧美精品| 久久久精品欧美日韩精品| 久久国内精品自在自线图片| 欧美另类亚洲清纯唯美| 人妻制服诱惑在线中文字幕| 国产午夜精品久久久久久一区二区三区 | 蜜臀久久99精品久久宅男| 国产精品,欧美在线| а√天堂www在线а√下载| 国产视频内射| 国产欧美日韩精品一区二区| 国产精品久久久久久亚洲av鲁大| 亚洲av美国av| 97热精品久久久久久| 99热6这里只有精品| 精品一区二区三区av网在线观看| 一个人看的www免费观看视频| 亚洲自拍偷在线| 欧美一区二区国产精品久久精品| 给我免费播放毛片高清在线观看| 国产av一区在线观看免费| 欧美激情久久久久久爽电影| 一级a爱片免费观看的视频| 欧美日韩乱码在线| 啦啦啦啦在线视频资源| 中文资源天堂在线| 日本五十路高清| 久久久精品欧美日韩精品| 99热6这里只有精品| 18+在线观看网站| 久久精品人妻少妇| a级毛片a级免费在线| 99热6这里只有精品| 久久久久国产精品人妻aⅴ院| 国产精品国产三级国产av玫瑰| 午夜激情欧美在线| 亚洲成人av在线免费| 色哟哟·www| 成熟少妇高潮喷水视频| 亚洲精品在线观看二区| 一本一本综合久久| 国产白丝娇喘喷水9色精品| 97在线视频观看| 内地一区二区视频在线| 亚洲乱码一区二区免费版| 国模一区二区三区四区视频| 国产亚洲精品综合一区在线观看| 国产精品免费一区二区三区在线| 久久99热6这里只有精品| 国产老妇女一区| a级毛片免费高清观看在线播放| 亚洲av不卡在线观看| 精品熟女少妇av免费看| av卡一久久| 亚洲av中文字字幕乱码综合| 99热只有精品国产| 欧美高清性xxxxhd video| 国产高潮美女av| 最近中文字幕高清免费大全6| 国产精品一区二区免费欧美| avwww免费| 国产麻豆成人av免费视频| 国产精品国产高清国产av| 色综合站精品国产| 观看免费一级毛片| 日本 av在线| 最近的中文字幕免费完整| h日本视频在线播放| 亚洲精品国产成人久久av| 国产aⅴ精品一区二区三区波| 国产精品久久久久久亚洲av鲁大| 精品欧美国产一区二区三| 不卡视频在线观看欧美| 久久久久久久午夜电影| 精品久久久噜噜| 一a级毛片在线观看| 能在线免费观看的黄片| 熟女电影av网| 色播亚洲综合网| 日韩欧美在线乱码| 亚洲内射少妇av| 国产精品久久电影中文字幕| 色在线成人网| 少妇熟女欧美另类| 国产老妇女一区| 日韩欧美免费精品| 国产精品野战在线观看| 九九久久精品国产亚洲av麻豆| 国产精品美女特级片免费视频播放器| 小说图片视频综合网站| 色播亚洲综合网| 亚洲成av人片在线播放无| 亚洲综合色惰| 晚上一个人看的免费电影| 男人舔奶头视频| 最近手机中文字幕大全| 成年版毛片免费区| 日本爱情动作片www.在线观看 | 一级毛片电影观看 | 亚洲精品国产av成人精品 | 成人无遮挡网站| 免费一级毛片在线播放高清视频| 久久中文看片网| a级毛色黄片| 日本黄色视频三级网站网址| 欧美在线一区亚洲| 国产精品乱码一区二三区的特点| 观看美女的网站| 淫妇啪啪啪对白视频| 五月伊人婷婷丁香| 最新在线观看一区二区三区| 国产三级中文精品| 99久久成人亚洲精品观看| 亚洲欧美日韩无卡精品| 好男人在线观看高清免费视频| 国产探花极品一区二区| 国产精品女同一区二区软件| 欧美高清成人免费视频www| 日韩三级伦理在线观看| 国产久久久一区二区三区| 亚洲精品一区av在线观看| 久久国内精品自在自线图片| 一级黄色大片毛片| 日日干狠狠操夜夜爽| 亚洲精华国产精华液的使用体验 | 欧美成人一区二区免费高清观看| 伊人久久精品亚洲午夜| 我要搜黄色片| 国产探花极品一区二区| 亚洲av第一区精品v没综合| 久久久欧美国产精品| 国产高清视频在线观看网站| 男女那种视频在线观看| 久久久久久久久久黄片| 尤物成人国产欧美一区二区三区| 激情 狠狠 欧美| 久久久久精品国产欧美久久久| 国产 一区精品| 麻豆精品久久久久久蜜桃| 国产精品久久久久久久电影| 久久午夜福利片| 人妻制服诱惑在线中文字幕| 男女之事视频高清在线观看| 欧美国产日韩亚洲一区| 天天一区二区日本电影三级| 亚洲av电影不卡..在线观看| 波野结衣二区三区在线| 成人三级黄色视频| 我要搜黄色片| 一本精品99久久精品77| 亚洲精品亚洲一区二区| 一区二区三区四区激情视频 | 亚洲欧美中文字幕日韩二区| 中国美白少妇内射xxxbb| 国产伦精品一区二区三区视频9| 99在线视频只有这里精品首页| 久久久久免费精品人妻一区二区| 男女做爰动态图高潮gif福利片| 热99re8久久精品国产| 黄色一级大片看看| 国产欧美日韩精品亚洲av| 国产v大片淫在线免费观看| 日韩人妻高清精品专区| 两个人视频免费观看高清| 久久久精品大字幕| 搡老岳熟女国产| 久久午夜福利片| 搞女人的毛片| 国产成人91sexporn| 狂野欧美白嫩少妇大欣赏| 伦理电影大哥的女人| 久久婷婷人人爽人人干人人爱| 一进一出抽搐动态| 欧美日韩一区二区视频在线观看视频在线 | 亚洲国产精品久久男人天堂| 两个人的视频大全免费| 综合色av麻豆| 欧美高清性xxxxhd video| 91久久精品电影网| 精品久久久久久久末码| 国产伦精品一区二区三区四那| 日产精品乱码卡一卡2卡三| 欧美成人一区二区免费高清观看| 精品久久久久久久人妻蜜臀av| 亚洲欧美日韩无卡精品| 国产淫片久久久久久久久| 欧美激情久久久久久爽电影| 香蕉av资源在线| 麻豆av噜噜一区二区三区| 久久精品91蜜桃| 舔av片在线| 极品教师在线视频| 12—13女人毛片做爰片一| 女人十人毛片免费观看3o分钟| 青春草视频在线免费观看| 国内少妇人妻偷人精品xxx网站| 最近中文字幕高清免费大全6| 可以在线观看毛片的网站| 日本 av在线| 国产视频内射| 小说图片视频综合网站| 69av精品久久久久久| 亚洲中文日韩欧美视频| 亚洲综合色惰| 亚洲最大成人手机在线| 国产精品不卡视频一区二区| 精品乱码久久久久久99久播| 成人二区视频| 国产一级毛片七仙女欲春2| 成人特级av手机在线观看| 国产又黄又爽又无遮挡在线| 2021天堂中文幕一二区在线观| 禁无遮挡网站| 国内精品一区二区在线观看| 色综合站精品国产| 国产在视频线在精品| 久久久久久九九精品二区国产| 久久久久国内视频| 国产黄片美女视频| 我要看日韩黄色一级片| 午夜精品一区二区三区免费看| 国产午夜精品久久久久久一区二区三区 | 久久久久久久午夜电影| 成熟少妇高潮喷水视频| 成人av一区二区三区在线看| 日日干狠狠操夜夜爽| 亚洲丝袜综合中文字幕| 色在线成人网| 日韩欧美精品免费久久| 免费av观看视频| 免费av毛片视频| 99热6这里只有精品| 国产老妇女一区| 乱人视频在线观看| 成人综合一区亚洲| av国产免费在线观看| 1024手机看黄色片| 亚洲国产精品合色在线| 长腿黑丝高跟| 蜜桃久久精品国产亚洲av| 日韩一本色道免费dvd| 99国产精品一区二区蜜桃av| 亚洲丝袜综合中文字幕| 久久这里只有精品中国| 久久久久性生活片| 国产精品久久久久久久电影| 欧美激情在线99| 欧美日本视频| АⅤ资源中文在线天堂| 日本在线视频免费播放| 国产 一区精品| 欧美成人免费av一区二区三区| 欧美日本亚洲视频在线播放| 亚洲精品456在线播放app| 精品久久久久久久久亚洲| 亚洲av不卡在线观看| 国产精品av视频在线免费观看| 乱码一卡2卡4卡精品| 99久久中文字幕三级久久日本| 国产在线精品亚洲第一网站| 久久精品综合一区二区三区| 欧美人与善性xxx| 男女之事视频高清在线观看| 老熟妇乱子伦视频在线观看| 一级毛片我不卡| av.在线天堂| 亚洲国产色片| 大型黄色视频在线免费观看| 中文字幕精品亚洲无线码一区| 99国产极品粉嫩在线观看| 久久久精品大字幕| 国产一区二区在线观看日韩| 别揉我奶头 嗯啊视频| 精品一区二区免费观看| 啦啦啦观看免费观看视频高清| 国产精品三级大全| 欧美人与善性xxx| 成人午夜高清在线视频| h日本视频在线播放| 久久九九热精品免费| 22中文网久久字幕| 干丝袜人妻中文字幕| 国产 一区精品| 欧美另类亚洲清纯唯美| 日韩三级伦理在线观看| 在线国产一区二区在线| 亚洲欧美日韩卡通动漫| 亚洲人成网站在线播放欧美日韩| 一级毛片我不卡| 波野结衣二区三区在线| 在线观看美女被高潮喷水网站| 成人美女网站在线观看视频| av在线观看视频网站免费| 一本一本综合久久| av天堂中文字幕网| 日本爱情动作片www.在线观看 | 看非洲黑人一级黄片| 最近在线观看免费完整版| 伊人久久精品亚洲午夜| 少妇的逼好多水| 色尼玛亚洲综合影院| 3wmmmm亚洲av在线观看| 久久久久性生活片| 精品一区二区三区人妻视频| 美女cb高潮喷水在线观看| 精品少妇黑人巨大在线播放 | 看免费成人av毛片| 精品久久国产蜜桃| 久久久精品大字幕| 美女黄网站色视频| 亚洲久久久久久中文字幕| 国产高潮美女av| 午夜福利视频1000在线观看| 色5月婷婷丁香| 精品久久久久久久久av| 亚洲国产精品久久男人天堂| 国产三级在线视频| 麻豆国产av国片精品| 欧美区成人在线视频| 国产黄色小视频在线观看| 中文字幕av在线有码专区| АⅤ资源中文在线天堂| 午夜久久久久精精品| 日韩中字成人| 亚洲精品456在线播放app| 最好的美女福利视频网| 狂野欧美白嫩少妇大欣赏| 悠悠久久av| 简卡轻食公司| 国产精品一区二区性色av| 男插女下体视频免费在线播放| h日本视频在线播放| 色播亚洲综合网| 免费看光身美女| 最近视频中文字幕2019在线8| 不卡视频在线观看欧美| 午夜福利视频1000在线观看| 日韩精品青青久久久久久| 午夜福利成人在线免费观看| 亚洲av美国av| 亚洲性久久影院| 在线观看午夜福利视频| 成人毛片a级毛片在线播放| 免费看av在线观看网站| 看非洲黑人一级黄片| 欧美一级a爱片免费观看看| 日韩强制内射视频| 波野结衣二区三区在线| 亚洲人与动物交配视频| 精品国产三级普通话版| 国产av麻豆久久久久久久| 免费大片18禁| 国产熟女欧美一区二区| 在线免费观看不下载黄p国产| 3wmmmm亚洲av在线观看| 日韩大尺度精品在线看网址| 成人漫画全彩无遮挡| 在线a可以看的网站| 亚洲成人久久爱视频| 亚洲成人中文字幕在线播放| 最近最新中文字幕大全电影3| 欧美性猛交╳xxx乱大交人| 色综合色国产| 亚洲精品影视一区二区三区av| 国产高清不卡午夜福利| 国产麻豆成人av免费视频| 日韩中字成人| 99精品在免费线老司机午夜| 在线观看美女被高潮喷水网站| 全区人妻精品视频| 国产单亲对白刺激| 深夜精品福利| 久久精品国产亚洲av涩爱 | av黄色大香蕉| 欧美性猛交黑人性爽| 国产精品一区二区三区四区免费观看 | 国产精品99久久久久久久久| 亚洲精品久久国产高清桃花| 日韩精品青青久久久久久| 欧美国产日韩亚洲一区| 综合色av麻豆| 波野结衣二区三区在线| 中文字幕精品亚洲无线码一区| 国产伦精品一区二区三区视频9| 国产激情偷乱视频一区二区| 天天躁日日操中文字幕| 欧美+亚洲+日韩+国产| 国产精品永久免费网站| 中文字幕熟女人妻在线| 欧美三级亚洲精品| 麻豆av噜噜一区二区三区| 亚洲国产欧洲综合997久久,| 国产精品乱码一区二三区的特点| 最近在线观看免费完整版| 久久国产乱子免费精品| 亚洲人成网站高清观看| 日本爱情动作片www.在线观看 | 99久国产av精品| 五月玫瑰六月丁香| 久久久久国产网址| 亚洲欧美日韩高清专用| 18禁黄网站禁片免费观看直播| 大又大粗又爽又黄少妇毛片口| 岛国在线免费视频观看| 在线观看av片永久免费下载| 直男gayav资源| 成人特级av手机在线观看| 在线天堂最新版资源| 成人二区视频| 成年女人永久免费观看视频| 99视频精品全部免费 在线| 女人被狂操c到高潮| 国产成人福利小说| 女同久久另类99精品国产91| 日本黄色片子视频| 有码 亚洲区| 免费av不卡在线播放| 亚洲一区高清亚洲精品| 亚洲欧美日韩无卡精品| 在线播放国产精品三级| 亚洲18禁久久av| 国产精品人妻久久久久久| 最近中文字幕高清免费大全6| 亚洲专区国产一区二区| 国产片特级美女逼逼视频| 亚洲精品在线观看二区| 女的被弄到高潮叫床怎么办| 九九久久精品国产亚洲av麻豆| 国产乱人偷精品视频| 九色成人免费人妻av| 18+在线观看网站| 国产高清有码在线观看视频| 成年版毛片免费区| 99久国产av精品国产电影|