• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Parallel Extraction of Marine Targets Applying OIDA Architecture

    2022-06-14 05:21:00LIULinLIWanwuZHANGJixianSUNYiandCUIYumeng
    Journal of Ocean University of China 2022年3期

    LIU Lin, LI Wanwu, *, ZHANG Jixian, SUN Yi, and CUI Yumeng

    Parallel Extraction of Marine Targets Applying OIDA Architecture

    LIU Lin1), LI Wanwu1), *, ZHANG Jixian2), SUN Yi1), and CUI Yumeng1)

    1),,266590,2),100830,

    Computing resources are one of the key factors restricting the extraction of marine targets by using deep learning. In order to increase computing speed and shorten the computing time, parallel distributed architecture is adopted to extract marine targets. The advantages of two distributed architectures, Parameter Server and Ring-allreduce architecture, are combined to design a parallel distributed architecture suitable for deep learning–Optimal Interleaved Distributed Architecture (OIDA). Three marine target extraction methods including OTD_StErf, OTD_Loglogistic and OTD_Sgmloglog are used to test OIDA, and a total of 18 experiments in 3 categories are carried out. The results show that OIDA architecture can meet the timeliness requirements of marine target extraction. The average speed of target parallel extraction with single-machine 8-core CPU is 5.75 times faster than that of single-machine single-core CPU, and the average speed with 5-machine 40-core CPU is 20.75 times faster.

    parallel computing; distributed architecture; deep learning; target extraction; PolSAR image

    1 Introduction

    The effect of marine target detection based on deep learning (DL) largely depends on the amount of training data and training scale. Therefore, with the help of parallel distributed architecture for model training, the training time of deep learning is reduced, which is beneficial to optimize the neural network model rapidly (Goyal, 2017), thus improving the speed and accuracy of marine target detection.

    Parallel architecture has been applied to related researches of target detection and extraction, and has achieved certain results. Parallel architecture and parallel tempering algorithm were used to conduct multi-target tracking, and the results show that it can make full use of parallel com- puting capabilities of GPU and improve the computational efficiency of the algorithm (Wen, 2017). In order to solve the timeliness problem of video processing, parallel architectures were designed to detect moving targets in the vi- deos, which significantly improved the computing speed (Peng., 2014; Lou., 2016; Zeng, 2017). Ling. (2016) proposed a parallel algorithm using sparse pulse- coupled neural network to detect moving targets based on Nvidia compute unified device architecture (CUDA), which improved the computational efficiency of the algorithm. Zhang (2017) changed traditional image processing algorithm based on CPU serial, and adopted GPU parallel algorithm to realize ship target detection in optical images. You (2016) derived fast detection algorithm for abnormal targets, proposed a parallel processing method under GPU architecture, and accelerated the calculation of hyperspectral data through CUDA, up to 33.2 times (Li, 2015). Different distributed architectures have been proposed to improve image processing speed (Hu., 2016; Quirita., 2016; Wu., 2016; Chen., 2017; Ye., 2017; Huang., 2018). And the parallel algorithm was enhanced from different angles to improve its applicabi- lity in deep learning (Aytekin, 2016; Mamidala., 2018; Cheng and Xu, 2019; Shen., 2019; Liu., 2020). Alqahtani and Demirbas (2019) established three parallel models of different system architectures including Parameter Server (PS), Peer-to-Peer, and Ring Allreduce (RA). Thao Nguyen. (2019) proposed two hierarchical distributed memory multi-lead AllReduce algorithms for GPU accelerated clusters optimization, and evaluated it on the discrete-event simulation simgrid. Bouzidi (2019) used MapReduce model to realize the parallel distributed algorithm based on the computing engine SPARK.

    On the basis of existing research results, the PS archi-tecture and the RA architecture are synthesized and a moreefficient parallel distributed architecture is designed according to the characteristics of the polarized Synthetic Aperture Radar (SAR) image and its deep learning data set.A single computer uses multi-core CPU/GPU to achieve pa- rallel computing, while multiple computers pre-load model training parameters through the parameter server to achieve uninterrupted distributed computing. In this way, parallel experiments are conducted on multiple extraction models of marine targets to test the performance of Optimal Interleaved Distributed Architecture (OIDA).

    2 Proposed Architecture

    2.1 Parallel Distributed Architecture

    Parallel distributed architectures mainly include two types: PS architecture and RA architecture.

    In PS architecture, the nodes of cluster are divided into two categories: PS and Worker. PS stores the parameters of the model, and Worker calculates the gradient of the parameters. In each iterative process, Worker obtains the parameters from PS, and then returns the calculated gradient to PS. PS aggregates the gradients returned from Worker, then updates the parameters and passes the new parameters to Worker. PS architecture is the most common-ly used distributed training architecture for DL. PS architecture using synchronous SGD is shown in Fig.1.

    The devices of ring-allreduce architecture are all Workers and form a ring, as shown in Fig.2, and there is no central node to aggregate the gradients calculated by all Workers. In the iterative process, each Worker calculates the gradient after completing its own mini-batch training, and passes it to the next Worker in the ring. Worker also re- ceives the gradient passed by the previous Worker at the same time. Each Worker updates the model parameters after receiving the gradients of all Workers except for itself. Compared with PS architecture, ring-allreduce architecture is bandwidth-optimized, which can make full use of the cha-racteristics of BP algorithm to reduce training time.

    2.2 OIDA Architecture

    The PS and RA architectures are integrated in the paper, and on this basis OIDA is designed, as shown in Fig.3. Its working principle is as follows: A single computer makes full use of multi-core CPU and GPU to achieve parallel computing, while multiple computers pre-load intermediate training parameters of model through the PS to achieve uninterrupted distributed computing. In an iterative pro- cess, each Worker of the child node calculates the gradientafter completing its own batch training, passes it to PServer, and reads the model from the model folder to continue training. The PServer of the child node passes the gradient and other parameters calculated by Worker at this node to the central node chief, receives the model parameters of the central node at the same time, updates the model in themodel folder of the node for Worker to call, and passes the model parameters to the upper and lower nodes connected to it. The upper and lower nodes check the model para- meters from the child nodes and compare with the model parameters of their own nodes to determine whether to up- date the models. PServer in the central node chief monitors and receives the model parameters passed by each child Node, passes them to Worker_DS in central node to test and evaluate model parameters, and updates the model parameters in the model folder for the child nodes to call.

    For OIDA deployment, first a cluster is created then job and task are assigned. For each task, host addresses are assigned, and a server is created. Cluster must be imported when sever is created so that each server knows which hosts are included in the cluster it is in, and then servers can communicate each other. Servers must be created on their own host. Once all servers are created on their respective hosts, the entire cluster is constructed, and servers be- tween clusters can communicate with each other. Each Ser- ver contains two components: Master and Worker. Master can mainly provide remote access (RPC protocol) to each device in the cluster. At the same time, it serves as a target for creating TF. Session. Worker can execute calculation sub- graphs with local equipments.

    Fig.1 Synchronous SGD training method in PS architecture.

    Fig.2 Ring-allreduce architecture.

    Fig.3 OIDA architecture.

    3 Models and Methods

    3.1 DL Model Construction

    A DL model OceanTDA9 is constructed to detect marine targets in this paper. OceanTDA9 contains 4 convolutional layers, 1 convolutional group and 3 fully connected layers, as shown in Fig.4. The first 4 convolution forms are same, and each group is as Convolution2D-relu-Dropout-Max- pool. The organization form of the intermediate convolution group is as (Convolution2D-relu-Dropout)*2-Max- pool. The last three are fully connected dense layers, among which the first two sets of dense are as Dense-relu-Drop- out, and the last fully connected dense layer only has dense. The kernel of each convolutional layer is a 3×3 small convolution kernel, and Dropout is set to 0.2. All pooling adopts maxpooling with a core size of 2×2 and a sliding step stride of 2.

    3.2 Target Extraction Method

    The marine target extraction methods are proposed in this paper, including Constant False Alarm Rate (CFAR) method based on the initial detection (OTD_StErf), marine target extraction method based on loglogistic (OTD_ Loglogistic), the method based on Adjoint Covariance Cor- rection Model (OTD_Sgmloglog) for complex sea conditions.

    Fig.4 DL model OceanTDA9 for marine target detection.

    The OTD_StErf method is to first adopt DL model OceanTDA9 for initial detection, and save the obtained suspected targets. 3×3 images, which contains 28×28 pixel,are constructed with the suspected target points as the cen- ter in turn. Then, the two-parameter CFAR method was used to perform n tests on the 3×3 window. The value ofis calculated by Eq. (1), and the critical pixel valueXis found by the target detection model (Eq. (2)). The pixels whose values are larger than the critical value in the 3×3 image window are determined as suspected targets.

    OTD_Loglogistic method first adopts the DL model OceanTDA9 for initial detection, and constructs a 3×3 image window with the obtained suspected target as the center, and saves the pixels that are not the targets in the window to the array. Calculate the frequency of each pixel inand draw a histogram, extract the effective pi-xels and corresponding frequencies from the smallest pixel to the largest pixel in the histogram and save them to the array2 and2 as fitting data. Call loglogistic probability density function(,,) (Eq.(3)) to fit2 and2, obtain the parameter values of (,)for the fitted curve, and evaluate the goodness of fit such as Chi-square and absolute value error.

    The principle of the OTD_Sgmloglog ocean target extraction method is similar to the above two methods, except that the calling model is different. It calls Adjoint Covariance Correction Model(,,) constructed by the author to fit2,2, as shown in Eq. (4), where?[0, 255],>0,>0, the modified parameter?[0, 1], andis the standard deviation of the fitted curve. Finally, the para- meter values of,andof the fitted curve are calculated, and the parameters indicating the goodness of fit such as Chi-square and absolute value error are calculated.

    4 Parallel Experiment

    In the paper, the experimental area for marine target parallel extraction is in the Bohai Sea, located in 37?07?– 40?56?N, 117?33?–122?08?E, and the polarized SAR data from Sentinel-1 in this area are used as the experimental data. Through experimental comparison, it is found that the VV (vertical transmission, vertical reception) polarizedSAR image from Sentinel-1 IW (Interferometric Wide Swath Mode) is more suitable for marine target extraction. So VV polarized SAR image is selected for experiments. After the VV polarized SAR image is preprocessed by the pro- cedure such as split, calibration, speckle filter, multi-look- ing, terrain correction, resampling and data conversion, the marine target detection data set is obtained, which is used as the data source of parallel distributed experiments for marine target extraction.

    The marine target detection DL model OceanTDA9 is adopted to learn and train the preprocessed PolSAR dataset according to the above parallel distributed architectures, and the parameters of the neural network model for marine target detection are obtained. The model parameters are called to detect the research area, and the obtained sus- pected targets are saved in the library files. The suspected targets detected in the area with the range of 39?52?– 40?13?N, 120?44?–121?03?E are marked by the magenta boxes in Fig.5, where the upper left corner is assigned as 0 rows and 0 columns, corresponding to 40?13?N and 120?44?E, the lower right corner is at 2715 rows and 2519 columns, corresponding to 39?52?N and 121?03?E. Similar situations in this paper will not be explained again. The OIDA architecture designed in this paper is adopted to do experiments on three methods, OTD_StErf, OTD_Log- logistic, and OTD_Sgmloglog, and extract distributed tar- gets in the research area.

    Fig.5 Suspected targets after initial detection (magenta box).

    4.1 Experiment Design

    The overall process of marine target extraction by using parallel distributed architecture is shown in Fig.6. Each computer in the cluster is configured with a 4-core 8- thread CPU with 32GB of memory, and two GPUs with 2GB of video memory. All computers form a Gigabit Eth- ernet. Each computer in the cluster is configured with network parameter models and PolSAR image data after DL to ensure that the necessary dynamic data with extre- mely low redundancy are transmitted in the network.

    After the software and hardware environment is set up, start the service, and designate a node as chief, which is responsible for managing each node, coordinating the ope- ration between each node, and completing the distribution of tasks, the collection of intermediate results and the inte- gration and visualization of results. After the task of chief is received by other Workers, the corresponding function is initialized according to the task requirements. The tar- get fitting parameter extraction function is used to com- plete the clustering of suspected targets, extract the position parameters such as the center coordinates and inclina-tion of the suspected targets, and extract the shape parameters such as length, width, and area, then use a straight line to fit the central axis and use an ellipse to fit the shape of the suspected target. Three target parameter extraction me-thods including OTD_StErf method, OTD_Loglogistic me- thod, and OTD_Sgmloglog method were used.

    After each Worker node in the cluster complete initial detection task assigned by chief, the suspected targets in the unit of DL sub-image (28 pixels×28 pixels) are saved to the corresponding folder, and chief is notified that the tasks have been completed. According to the load condition of Workers in the cluster, the collected suspected targets are grouped and distributed to the corresponding Workers by chief. After the task of chief is received by Worker nodes, the target parameters are called to extract the function. According to the pre-set maximum and minimum length and aspect ratio, the target parameters that meet the conditions

    are extracted and saved to the corresponding folder, and Chief is notified that the tasks have been completed. The target parameters extracted by each Worker in the cluster are collected by chief and dynamically visualized after integration.

    Fig.6 Marine target extraction procedure by using parallel distributed architecture.

    4.2 Experiment on OTD_StErf

    For comparison, single-machine single-core device is used to extract marine target parameters for 35 suspected targets detected initially by OTD_StErf. It takes 381.45s to extract 22 targets. The time sequence is shown in Fig.7, and each target extraction task takes approximately 6s. In order to ensure the display effect and data storage, each target is set to delay 5s after displayed and then turned off.

    The OIDA designed in this paper is used for marine target extraction experiment based on OTD_StErf. The times for every CPU in every node of the experiment to display the suspected targets and extract targets are shown in Fig.8,where Fig.8(a) is the parallel computing with single-ma- chine 8-core CPU. The 35 suspected targets are allocated randomly to 0–7-core CPU, 4–5 tasks for one core. After each CPU receives the tasks, it will display positions of suspected targets in the research marine area in turn, and extract targets according to the target size and aspect ratio specified by the user.From Fig.8(a), it can be found that the suspected target detection takes approximately 6–7s, and each target extraction takes approximately 5–7s.

    Fig.8(b) is the parallel computing with dual-machine 16-core CPU, and 1–3 tasks are randomly allocated to 0–15 core CPU. The times for suspected target detection and each target extraction are approximately 6–7s. The first to complete the task is the 5-core and 6-core CPU, which are allocated 2 tasks and take 10s. The last one to complete the task is 1-core CPU, which takes approximately 33s. The CPU is assigned a total of 3 tasks, that is, 3 suspected tar- gets need to be detected, of which 2 suspected targets meet the conditions. The extraction times of these 2 targets are 10s and 25s, respectively.

    Fig.8(c) is the parallel computing with 3-machine 24- core CPU, and each CPU is randomly assigned 1–2 tasks. The first CPU that completes tasks takes 3s, and the last takes approximately 27s. Fig.8(d) is the parallel computing with 4-machine 32-core CPU, 1–2 tasks are randomly allocated to each CPU. The first CPU that completes tasks takes 1s, and the last takes approximately 23s. Fig.8(e) is the parallel computing with 5-machine 40-core CPU. Since there are 35 tasks, 35 CPU cores are involved in the calculation, and 1 task is randomly allocated to one CPU. The first CPU that completes tasks takes 1s, and the last takes approximately 9s. The 0–7 core CPUs in the calculationshown in Fig.8(b–e) are the chief CPUs. Since chief is responsible for assigning tasks to other Workers, the time for the chief CPU to display the first suspected target is when the program runs for 3s, and the time for other Workers to display the first suspected target is when the program runs for 1s, 0 or 3s, 1 or 2s, 1 or 2s.

    The CPU task execution status of each node is shown in Table 1. In single-machine 8-core CPU parallel architecture, each CPU takes approximately 53.77s, all CPUs take a total of 430.23s, and the total time to complete the task is 67.07s, which is 17.58% of the time consumed by a single machine with one core,., 82.49% faster. In 5- machine 40-core CPU parallel architecture, the cluster com- posed of CPUs take 17.66s to complete the task, which is 4.62% of the time consumed by a single machine, and 95.38% faster. The CPU execution status in other clusters are shown in Table 1.

    Fig.7 The display time of target parameter extraction when OTD_StErf method is carried out under single-machine single- core condition. The suspected targets (Starget) are represented by the red dot ‘?’, and the extracted targets (Etarget) are represented by the blue plus ‘+’. Here in after the same.

    Fig.8 Parallel extraction CPU-t diagram for OTD_StErf.

    Table 1 Parallel extraction CPU status for OTD_StErf

    4.3 Experiment on OTD_Loglogistic

    Single-machine single-core architecture is adopted to extract marine target parameters for 35 suspected targets, detected initially after DL, by OTD_Loglogistic, which takes 389.02s and extracts 22 targets. As shown in Fig.9, each extraction takes approximately 6–9s, and the first sus- pected target is displayed when the program runs for 1s.

    The OIDA framework designed in this paper is used to extract marine target parameters by OTD_Loglogistic me- thod. The times that the CPU of each core at each node dis- plays the suspected targets and extracts them are shown in Fig.10. Among them, Fig.10(a) is the parallel computing with single-machine 8-core CPU. The 35 suspected targets are randomly allocated to 0–7 core CPU, 4–5 tasks for one core. The first CPU that completes tasks takes 22 s, and the last takes approximately 60s. Fig.10(b) is the parallel computing with dual-machine 16-core CPU. 1–3 tasks are randomly allocated to 0–15 core CPU. The first CPU that completes tasks takes 3s, and the last takes approximately 33s. Fig.10(c) is the parallel computing with 3-machine 24-core CPU. 1–2 tasks are randomly allocat- ed to each CPU. The first CPU that completes tasks takes 7s, and the last takes approximately 25s. Fig.10(d) is the parallel computing with 4-machine 32-core CPU. 1–2 tasks are randomly allocated to each CPU. The first CPU that completes tasks takes 2s, and the last takes approximately 24s. Fig.10(e) is the parallel computing with 5-machine 40-core CPU. 1 task is randomly allocated to each CPU. The first CPU that completes tasks takes 1s, and the last takes approximately 11s. In Figs.10b–e, the first suspect- ed target is displayed by the chief CPU when the program runs for 3s or 2s, and the first suspected target is display- ed by the other working nodes when the program runs for 2s, 1s, 2s, 1s or 2s, respectively.

    In the parallel extraction process of marine targets by single-machine multi-core/multi-machine multi-core OTD_ Loglogistic method, the CPU task execution status at each node is shown in Table 2. When the single-machine 8- core CPU is used, each CPU takes approximately 54.89s, all CPUs take a total of 439.18s, and the total time to complete the task is 68.40s, which is 17.58% of the time consumed by a single machine with single core, and 82.42% faster. The cluster composed of 5-machine 40-core CPU takes 18.91s to complete the task, which is 4.86% of the time consumed by a single machine with single core, and 95.14% faster. The CPU execution status in other clusters are shown in Table 2.

    4.4 Experiment on OTD_Sgmloglog

    An OTD_Sgmloglog method for marine target extraction in complex sea conditions is proposed by the author. The designed OIDA architecture is tested on this propos- ed method and the results are compared with the single- machine single-core mode.

    Single-machine single-core is used to extract marine target parameters by OTD_Sgmloglog method for 35 sus- pected targets detected initially after DL. It takes 389.52s to extract 22 targets. As shown in Fig.11, each target extraction takes approximately 6–7s, and the first suspected target is displayed when the program runs for 1s.

    Fig.9 Display time of target parameter extraction for single-machine single-core OTD_Loglogistic method.

    Fig.10 Parallel extraction CPU-t diagram for OTD_ Loglogistic.

    Table 2 Parallel extraction CPU status for OTD_Loglogistic method

    The designed OIDA architecture is used to extract marine target parameters by OTD_ Sgmloglog method. The times for each CPU at each node to display the suspected targets and extract targets are shown in Fig.12. Fig.12(a) is the parallel computing with single-machine 8-core CPU. The 35 suspected targets are randomly allocated to 0–7 core CPU, 4–5 tasks for one core. The first CPU that com- pletes tasks takes 23s, and the last takes approximately 60s. Fig.12(b) is the parallel computing with dual-machine 16-core CPU. 1–3 tasks are randomly allocated to 0–15 core CPU. The first CPU that completes tasks takes 3s, and the last takes approximately 33s. Fig.12(c) is the paral- lel computing with 3-machine 24-core CPU. 1–2 tasks are randomly allocated to each CPU. The first CPU that com- pletes tasks takes 3s, and the last takes approximately 25s. Fig.12(d) is the parallel computing with 4-machine 32- core CPU. 1–2 tasks are randomly allocated to each CPU. The first CPU that completes tasks takes 1s, and the last takes approximately 24s. Fig.12(e) is the parallel computing with 5-machine 40-core CPU. 1 task is randomly allocated to each CPU. The first CPU that completes tasks takes 1s, and the last takes approximately 12s. The 0–7 core CPU in Figs.12b–e is the chief CPU and the first suspected target is displayed when the program runs for 3 or 2s. The first suspected target for other working nodes is displayed when the program runs for 2s, 1s or 2 s, 1s or 2s, 1s or 2s, respectively.

    In the parallel extraction of marine target parameters by single-machine multi-core/multi-machine multi-core OTD_ Sgmloglog method, the CPU status of each node is shown in Table 3. Each CPU of the single-machine 8-core takes approximately 54.23s, all CPUs take a total of 433.85s, and the total time to complete the task is 67.40s, which is 17.30% of the time consumed by a single machine with single core, and 82.70% faster. The cluster composed of 5-machine 40-core CPU takes 20.17s to complete the task, which is 5.18% of the time consumed by a single machine with single core, and 94.82% faster. The CPU execution status in the other clusters is shown in Table 3.

    Fig.11 CPU-t diagram of parameter extraction for single-machine single-core OTD_ Sgmloglog method.

    Fig.12 Parallel extraction CPU-t diagram for OTD_ Sgmloglog method.

    Table 3 Parallel extraction CPU status for OTD_Sgmloglog method

    4.5 Results for Marine Target Extraction

    The above method is used to extract target parameters from the relevant SAR data in the research marine area, and the parameters of target position and shape are fitted by ellipses. The fitting results are shown in Fig.13, and the area marked by blue box is enlarged in Fig.14. The detailed ellipse fitting parameters of the extracted targets are listed in Table 4.

    5 Discussion and Conclusions

    The OIDA parallel distributed architecture is designed in the paper, and comparison experiments are conducted on three extraction methods of marine targets: OTD_ StErf, OTD_Loglogistic, and OTD_Sgmloglog. Each me- thod is applied under 6 deployments of single-machine single-core CPU, single-machine 8-core CPU, dual-ma-chine 16-core CPU, 3-machine 24-core CPU, 4-machine 32-core CPU, and 5-machine 40-core CPU, including 18 experiments. Experimental results show that the parallel distributed OIDA architecture which is designed in this paper is efficient in the applications of marine target extraction. The average speed of parallel target extraction for a single-machine 8-core CPU is 5.75 times the speed of a single-machine single-core CPU, where the highest is 5.84 times, and the lowest is 5.78 times. The average speed of parallel target extraction for a 5-machine 40-core CPU is 20.75 times the speed of a single-machine single-core CPU, where the highest is 21.53 times and the lowest is 19.31 times. The parallel distributed OIDA architecture shortens the time required for the operation of the marine target extraction method. The average response time per thousand square kilometers is approximately 2s, with the fastest response time of 1s, and the slowest response time of 3s. The designed parallel distributed architecture OIDA realizes the high-precision rapid extraction of target features.

    Fig.13 Marine target extraction results.

    Fig.14 Local enlargement of marine target extraction results.

    Acknowledgements

    The study was supported by the Natural Science Foundation of Shandong Province (No. ZR2019MD034), and the Education Reform Project of Shandong Province (No. M2020266). This study was supported by data from European Space Agency (ESA) and the Institute of Remote Sensing and Digital Earth, Chinese Academy of Sciences.

    Alqahtani, S., and Demirbas, M., 2019. Performance analysis and comparison of distributed machine learning systems.: 1909.02061.

    Aytekin, A., Feyzmahdavian, H. R., and Johansson, M., 2016. Analysis and implementation of an asynchronous optimization algorithm for the parameter server.: 1610.05507.

    Bouzidi, S., 2019. Parallel and distributed implementation on SPARK of a spectral-spatial classifier for hyperspectral images., 13 (3): 034501.

    Chen, H., Wei, A., and Zhang, Y., 2017. Three-level parallel-set partitioning in hierarchical trees coding based on the collaborative CPU and GPU for remote sensing images compression., 11 (4): 045015.

    Cheng, Z., and Xu, Z., 2019. Bandwidth reduction using importance weighted pruning on ring allreduce.: 1901.01544.

    Goyal, P., Dollár, P., Girshick, R., and Noordhuis, P., 2017. Accurate, large minibatch SGD: Training imagenet in 1hour.: 1706.02677.

    Hu, H., Shu, H., Hu, Z., and Xu, J., 2016. Using compute unified device architecture-enabled graphic processing unit to accelerate fast fourier transform-based regression kriging interpolation on a MODIS land surface temperature image., 10 (2): 026036.

    Huang, Y., Jin, T., Wu, Y., Cai, Z., and Cheng, J., 2018. Flexps: Flexible parallelism control in parameter server architecture., 11 (5): 566-579.

    Li, T., 2015. Parallel optimization and application research on moving object detection and recognition algorithms. Master thesis.National University of Defense Technology.

    Ling, B., Deng, Y., and Yu, S. B., 2016. Processing for accelerated sparse PCNN moving target detection algorithm with CUDA., 37 (12): 3300- 3305.

    Liu, Y. H., Zhou, J., Qi, W. H., Li, X. L., Gross, L., Shao, Q.,., 2020. ARC-Net: An efficient network for building extraction from high resolution aerial images., 8: 154997- 155010, DOI: 10.1109/ACCESS.2020.3015701.

    Lou, X. H., Guo, C. S., Song, S. L., and Qi, L. Q., 2016. Parallel implementation of video moving object detecion algorithm based on CUDA., 36 (3): 23-26.

    Mamidala, A. R., Kollias, G., Ward, C., and Artico, F., 2018. MXNET-MPI: Embedding MPI parallelism in parameter servertask model for scaling deep learning.: 18 01.03855.

    Peng, B., Zhang, C. Y., Zheng, S. B., and Tian, G., 2014. Multi- level parallel optimization of moving object detecion and fea- ture extraction algorithm., 38 (13): 173-177.

    Quirita, V., da Costa, G., Happ, P., Feitosa, P., da Silva Ferreyra R., Oliveira, D.,., 2016. A new cloud computing architecture for the classification of remote sensing data., 10 (2): 409-416.

    Shen, S., Xu, L., Liu, J., Liang, X., and Cheng, Y., 2019. Faster distributed deep net training: Computation and communication decoupled stochastic gradient descent.: 1906.12043,

    Thao Nguyen, T., Wahib, M., and Takano, R., 2019. Efficient MPI-allreduce for large-scale deep learning on GPU-clusters.: e5574.

    Wen, S., 2017. The parallel design and implementation of MCMC multi-object tracking algorithm. Master thesis. Xidian Univer- sity.

    Wu, Z., Li, Y., Plaza, A., Li, J., Xiao, F., and Wei, Z., 2016. Pa- rallel and distributed dimensionality reduction of hyperspectral data on cloud computing architectures., 9 (6): 2270-2278.

    Ye, L. H., Wang, L., Sun, Y. X., Zhao, L. P., and Wei, Y. W., 2017. Parallel multi-stage features fusion of deep convolutional neural networks for aerial scene classification., 9: 3, 294-303.

    You, W., 2016. Research on hyperspectral remote sensing target detection parallel processing. Master thesis. Harbin Engineering University.

    Zeng, T., 2017. Research on parallel algorithm for moving object contour extraction in video sequences. Master thesis. Cen- tral China Normal University.

    Zhang, F., 2017. Target recognition and parallel acceleration with GPU in marine remote sensing image. Master thesis. Changchun Institute of Optics, Fine Mechanics and Physics,Chinese Academy of Sciences.

    J. Ocean Univ. China(Oceanic and Coastal Sea Research)

    https://doi.org/10.1007/s11802-022-4839-z

    ISSN 1672-5182, 2022 21 (3): 737-747

    (November 16, 2020;

    December 24, 2020;

    January 13, 2021)

    ? Ocean University of China, Science Press and Springer-Verlag GmbH Germany 2022

    Corresponding author. E-mail: liwanwuqd@126.com

    (Edited by Chen Wenwen)

    一a级毛片在线观看| 成在线人永久免费视频| 激情在线观看视频在线高清| a级毛片黄视频| 大香蕉久久成人网| 一二三四社区在线视频社区8| 欧美日韩亚洲高清精品| 精品高清国产在线一区| 欧美中文日本在线观看视频| 超色免费av| 久久久水蜜桃国产精品网| 日本五十路高清| 亚洲欧美精品综合一区二区三区| 高清在线国产一区| av福利片在线| 嫩草影院精品99| 久久亚洲精品不卡| 18禁观看日本| 又黄又爽又免费观看的视频| 精品国内亚洲2022精品成人| 日本wwww免费看| 黄色片一级片一级黄色片| 久久人妻福利社区极品人妻图片| 精品久久久久久,| 国产精品乱码一区二三区的特点 | 在线天堂中文资源库| 免费少妇av软件| 久久久久久免费高清国产稀缺| 搡老乐熟女国产| 极品人妻少妇av视频| 一区二区三区精品91| 天堂中文最新版在线下载| 在线永久观看黄色视频| 在线观看一区二区三区激情| 亚洲成人免费电影在线观看| 国产亚洲精品综合一区在线观看 | 99久久国产精品久久久| 国产亚洲精品久久久久5区| 精品人妻在线不人妻| 999精品在线视频| 中文字幕精品免费在线观看视频| 午夜成年电影在线免费观看| 91国产中文字幕| 淫秽高清视频在线观看| 天天影视国产精品| 女人被躁到高潮嗷嗷叫费观| 亚洲熟女毛片儿| 亚洲色图av天堂| www.熟女人妻精品国产| 国产精品亚洲av一区麻豆| 十八禁人妻一区二区| 亚洲精品国产精品久久久不卡| 国产真人三级小视频在线观看| 亚洲国产欧美日韩在线播放| 搡老熟女国产l中国老女人| 一级片'在线观看视频| 成人手机av| 叶爱在线成人免费视频播放| 国产免费av片在线观看野外av| 国产欧美日韩一区二区三区在线| 国产熟女xx| 在线播放国产精品三级| aaaaa片日本免费| 国产精品久久久久久人妻精品电影| 色综合站精品国产| 国产一区二区激情短视频| 91麻豆精品激情在线观看国产 | e午夜精品久久久久久久| 成人手机av| www.www免费av| 久久精品91无色码中文字幕| 中亚洲国语对白在线视频| 午夜亚洲福利在线播放| 久久精品人人爽人人爽视色| 日韩三级视频一区二区三区| 99热只有精品国产| 精品人妻在线不人妻| 男女下面进入的视频免费午夜 | av超薄肉色丝袜交足视频| 久久久久久久精品吃奶| 一级片免费观看大全| 欧美在线一区亚洲| 国产成人精品久久二区二区91| 亚洲性夜色夜夜综合| 90打野战视频偷拍视频| 成在线人永久免费视频| 日韩欧美免费精品| 中文字幕另类日韩欧美亚洲嫩草| 91av网站免费观看| 国产精品日韩av在线免费观看 | 亚洲视频免费观看视频| 国产不卡一卡二| 三上悠亚av全集在线观看| 一二三四在线观看免费中文在| 欧美 亚洲 国产 日韩一| 美女福利国产在线| 91麻豆av在线| 精品午夜福利视频在线观看一区| 国产国语露脸激情在线看| 麻豆国产av国片精品| 麻豆国产av国片精品| 麻豆国产av国片精品| 亚洲色图 男人天堂 中文字幕| 亚洲五月天丁香| 人成视频在线观看免费观看| 啪啪无遮挡十八禁网站| 国产成人av教育| 成熟少妇高潮喷水视频| 国产三级在线视频| 精品国产亚洲在线| 亚洲精品在线观看二区| 国产一区在线观看成人免费| 国产亚洲欧美在线一区二区| 一级片免费观看大全| 一级片'在线观看视频| 99久久99久久久精品蜜桃| 看黄色毛片网站| 久久久久久久久中文| 91九色精品人成在线观看| 色婷婷久久久亚洲欧美| 色老头精品视频在线观看| 亚洲欧美精品综合一区二区三区| 女同久久另类99精品国产91| 女同久久另类99精品国产91| 久久欧美精品欧美久久欧美| 天天影视国产精品| 19禁男女啪啪无遮挡网站| 午夜福利免费观看在线| 满18在线观看网站| 日韩av在线大香蕉| 一级片免费观看大全| 一进一出抽搐动态| 日韩精品青青久久久久久| 国产精品二区激情视频| 18禁裸乳无遮挡免费网站照片 | 午夜影院日韩av| 国产午夜精品久久久久久| 如日韩欧美国产精品一区二区三区| 精品无人区乱码1区二区| 两人在一起打扑克的视频| 制服人妻中文乱码| 日韩免费高清中文字幕av| 亚洲自偷自拍图片 自拍| 亚洲色图av天堂| 久久久久久人人人人人| 亚洲国产精品sss在线观看 | 亚洲av第一区精品v没综合| 狠狠狠狠99中文字幕| 黄片大片在线免费观看| 一级作爱视频免费观看| 高清毛片免费观看视频网站 | 亚洲男人的天堂狠狠| 成人18禁高潮啪啪吃奶动态图| 成人影院久久| 一级毛片精品| 脱女人内裤的视频| 超碰成人久久| 欧美+亚洲+日韩+国产| 美女午夜性视频免费| 国产精品免费一区二区三区在线| 国产一区二区三区综合在线观看| 国产一区二区在线av高清观看| 国产一区二区三区综合在线观看| 嫩草影院精品99| 18禁美女被吸乳视频| 高清黄色对白视频在线免费看| 色精品久久人妻99蜜桃| 真人一进一出gif抽搐免费| 在线观看免费午夜福利视频| 久久久国产成人精品二区 | 亚洲精品中文字幕在线视频| 国产高清视频在线播放一区| 亚洲全国av大片| 欧美精品啪啪一区二区三区| 一级毛片精品| 老司机亚洲免费影院| 国产精品久久久久成人av| 亚洲精品中文字幕在线视频| 精品久久久久久,| 在线永久观看黄色视频| 欧美日韩中文字幕国产精品一区二区三区 | 人人妻人人爽人人添夜夜欢视频| 国产精品爽爽va在线观看网站 | 女人被狂操c到高潮| 欧洲精品卡2卡3卡4卡5卡区| 我的亚洲天堂| 在线国产一区二区在线| 亚洲一区中文字幕在线| 日韩精品青青久久久久久| avwww免费| 国产三级黄色录像| 久久亚洲真实| 国产欧美日韩综合在线一区二区| 国产激情久久老熟女| 日日爽夜夜爽网站| 亚洲av五月六月丁香网| 国产成人欧美| 韩国精品一区二区三区| 黄色女人牲交| 91av网站免费观看| 真人一进一出gif抽搐免费| 首页视频小说图片口味搜索| 久久久久精品国产欧美久久久| 看片在线看免费视频| 国产乱人伦免费视频| 嫩草影视91久久| 亚洲精品一二三| 国产精品爽爽va在线观看网站 | 成年女人毛片免费观看观看9| 国产精品永久免费网站| 欧美日韩瑟瑟在线播放| 老汉色∧v一级毛片| 亚洲伊人色综图| 国产91精品成人一区二区三区| 黄片小视频在线播放| 亚洲七黄色美女视频| 久久香蕉激情| 久久久国产成人精品二区 | 久久狼人影院| 色综合站精品国产| 99精国产麻豆久久婷婷| а√天堂www在线а√下载| 成人黄色视频免费在线看| 首页视频小说图片口味搜索| 黑人欧美特级aaaaaa片| 国产精品一区二区精品视频观看| 波多野结衣av一区二区av| 黄频高清免费视频| 黑人猛操日本美女一级片| 亚洲精品国产区一区二| 欧美激情极品国产一区二区三区| 一二三四在线观看免费中文在| 欧美不卡视频在线免费观看 | 国产成人欧美在线观看| 精品欧美一区二区三区在线| 国产亚洲欧美精品永久| 亚洲中文日韩欧美视频| 国产精品亚洲一级av第二区| 韩国精品一区二区三区| 在线观看一区二区三区| 在线看a的网站| 99热国产这里只有精品6| 成人三级做爰电影| 人人澡人人妻人| 丰满饥渴人妻一区二区三| 1024香蕉在线观看| 老鸭窝网址在线观看| 国产国语露脸激情在线看| 搡老乐熟女国产| 一本综合久久免费| 亚洲欧美日韩另类电影网站| 亚洲成人免费av在线播放| 亚洲精品一二三| 香蕉丝袜av| 国产精华一区二区三区| 99久久精品国产亚洲精品| 伦理电影免费视频| 亚洲欧美日韩高清在线视频| 欧美日韩国产mv在线观看视频| 成年版毛片免费区| 精品电影一区二区在线| 国产亚洲av高清不卡| 亚洲精品av麻豆狂野| 亚洲一区中文字幕在线| 999久久久国产精品视频| 国产不卡一卡二| 露出奶头的视频| 亚洲黑人精品在线| 一区在线观看完整版| 国产精品亚洲av一区麻豆| 9色porny在线观看| 咕卡用的链子| 亚洲中文字幕日韩| 人人妻,人人澡人人爽秒播| 精品人妻在线不人妻| 五月开心婷婷网| 久久精品91无色码中文字幕| 久久久精品国产亚洲av高清涩受| 国产av一区在线观看免费| 午夜日韩欧美国产| 国产亚洲精品久久久久久毛片| 波多野结衣av一区二区av| 亚洲,欧美精品.| 午夜福利免费观看在线| 这个男人来自地球电影免费观看| 国产麻豆69| av天堂久久9| netflix在线观看网站| 黄网站色视频无遮挡免费观看| 午夜影院日韩av| 国产欧美日韩精品亚洲av| 天堂中文最新版在线下载| 久久香蕉激情| 午夜成年电影在线免费观看| 亚洲自偷自拍图片 自拍| 天天躁夜夜躁狠狠躁躁| 少妇被粗大的猛进出69影院| 日韩视频一区二区在线观看| 十八禁人妻一区二区| 中文字幕色久视频| 欧美最黄视频在线播放免费 | 欧美 亚洲 国产 日韩一| 女人被躁到高潮嗷嗷叫费观| 午夜福利一区二区在线看| 国产av精品麻豆| 久久久国产成人免费| 国内久久婷婷六月综合欲色啪| 国产精品98久久久久久宅男小说| 神马国产精品三级电影在线观看 | av视频免费观看在线观看| 日本精品一区二区三区蜜桃| 精品久久久久久久毛片微露脸| 国产色视频综合| 国产高清videossex| 国产熟女xx| 欧美一区二区精品小视频在线| 午夜免费观看网址| 18禁美女被吸乳视频| 精品一区二区三区四区五区乱码| 两个人看的免费小视频| 欧美精品亚洲一区二区| 国产av在哪里看| 亚洲男人天堂网一区| 国产精品爽爽va在线观看网站 | 成人18禁在线播放| 一区二区三区激情视频| 长腿黑丝高跟| 欧美日韩视频精品一区| 久久青草综合色| 久久久久久人人人人人| 亚洲久久久国产精品| 亚洲人成电影观看| www.精华液| 中文亚洲av片在线观看爽| 午夜福利,免费看| 十八禁人妻一区二区| www.熟女人妻精品国产| 亚洲欧美一区二区三区久久| 超碰97精品在线观看| 亚洲一区二区三区色噜噜 | 亚洲av电影在线进入| 波多野结衣一区麻豆| xxxhd国产人妻xxx| 亚洲国产欧美日韩在线播放| 日本黄色日本黄色录像| 欧美色视频一区免费| 午夜免费鲁丝| 成人av一区二区三区在线看| 国产精品日韩av在线免费观看 | 午夜精品在线福利| 久久午夜综合久久蜜桃| 久久香蕉精品热| 一进一出好大好爽视频| 久久性视频一级片| 人妻丰满熟妇av一区二区三区| 激情视频va一区二区三区| 欧美老熟妇乱子伦牲交| 久9热在线精品视频| 啦啦啦在线免费观看视频4| 悠悠久久av| 久久香蕉国产精品| 日本撒尿小便嘘嘘汇集6| 美国免费a级毛片| 亚洲国产看品久久| 日韩高清综合在线| 一本综合久久免费| 亚洲成人国产一区在线观看| 久久久久久亚洲精品国产蜜桃av| 动漫黄色视频在线观看| 国产成人精品无人区| 我的亚洲天堂| 两个人免费观看高清视频| 两人在一起打扑克的视频| 亚洲精品中文字幕在线视频| 免费观看人在逋| 黄色a级毛片大全视频| 在线观看一区二区三区| 99香蕉大伊视频| 韩国av一区二区三区四区| 天天影视国产精品| 久久久久九九精品影院| 丝袜美足系列| 亚洲精品一区av在线观看| 十八禁人妻一区二区| 亚洲国产欧美网| 午夜福利欧美成人| 日韩大码丰满熟妇| 亚洲在线自拍视频| 乱人伦中国视频| 欧美成人性av电影在线观看| 精品欧美一区二区三区在线| 国产精品98久久久久久宅男小说| 日本精品一区二区三区蜜桃| 国产欧美日韩综合在线一区二区| 女人爽到高潮嗷嗷叫在线视频| 婷婷六月久久综合丁香| 18禁裸乳无遮挡免费网站照片 | 亚洲va日本ⅴa欧美va伊人久久| 精品国产一区二区三区四区第35| 国产伦人伦偷精品视频| 午夜成年电影在线免费观看| 亚洲中文字幕日韩| 国产一区二区在线av高清观看| 夜夜夜夜夜久久久久| 又黄又爽又免费观看的视频| 久久中文看片网| 女警被强在线播放| 9热在线视频观看99| 中文字幕熟女人妻在线| 欧美黑人欧美精品刺激| 成人午夜高清在线视频| 欧洲精品卡2卡3卡4卡5卡区| 国产综合懂色| av在线蜜桃| 欧美性猛交黑人性爽| 桃色一区二区三区在线观看| 免费看光身美女| 亚洲片人在线观看| 一本一本综合久久| 亚洲av中文字字幕乱码综合| 免费看a级黄色片| 乱人视频在线观看| 中文字幕免费在线视频6| 久久精品91蜜桃| 成人特级av手机在线观看| 可以在线观看的亚洲视频| 国产精品一区二区免费欧美| 久久热精品热| 男女之事视频高清在线观看| 欧美高清性xxxxhd video| 国产一区二区三区视频了| 嫩草影院精品99| 欧美高清性xxxxhd video| 日本黄大片高清| a级毛片免费高清观看在线播放| 成年女人毛片免费观看观看9| h日本视频在线播放| 天堂动漫精品| 在线免费观看不下载黄p国产 | 在线免费观看不下载黄p国产 | 国产色婷婷99| 亚洲真实伦在线观看| 男女视频在线观看网站免费| 男女那种视频在线观看| 国产乱人视频| 嫩草影院入口| 欧美潮喷喷水| 美女大奶头视频| 中亚洲国语对白在线视频| 啦啦啦韩国在线观看视频| 午夜福利18| 日韩欧美精品免费久久 | 男女床上黄色一级片免费看| 免费在线观看成人毛片| 老熟妇乱子伦视频在线观看| 少妇裸体淫交视频免费看高清| 99久国产av精品| 日韩中文字幕欧美一区二区| 成人特级av手机在线观看| 精品午夜福利视频在线观看一区| 欧美最新免费一区二区三区 | 男女视频在线观看网站免费| 成人一区二区视频在线观看| 2021天堂中文幕一二区在线观| 性插视频无遮挡在线免费观看| av在线老鸭窝| 大型黄色视频在线免费观看| 69av精品久久久久久| 亚洲人成网站高清观看| 中文字幕av成人在线电影| 欧美zozozo另类| 亚洲av成人精品一区久久| 欧美日本亚洲视频在线播放| 高清毛片免费观看视频网站| 日日摸夜夜添夜夜添av毛片 | 99国产极品粉嫩在线观看| 欧美黑人巨大hd| 精品久久久久久久人妻蜜臀av| av中文乱码字幕在线| 99久久无色码亚洲精品果冻| 午夜视频国产福利| 九九久久精品国产亚洲av麻豆| 国产人妻一区二区三区在| 黄色配什么色好看| 国产一区二区在线av高清观看| 亚洲欧美精品综合久久99| 我的老师免费观看完整版| 国产精品久久久久久人妻精品电影| 国产色婷婷99| 丁香欧美五月| 欧美色视频一区免费| 国产毛片a区久久久久| 精品久久久久久久久av| 中文字幕免费在线视频6| 久久热精品热| 久久久久国内视频| 好看av亚洲va欧美ⅴa在| 在线免费观看的www视频| 99久久精品国产亚洲精品| 999久久久精品免费观看国产| 男女之事视频高清在线观看| 精品国内亚洲2022精品成人| 久久精品夜夜夜夜夜久久蜜豆| 美女 人体艺术 gogo| 午夜两性在线视频| 噜噜噜噜噜久久久久久91| 午夜久久久久精精品| 久久人人精品亚洲av| 久久午夜亚洲精品久久| 欧美乱色亚洲激情| 变态另类丝袜制服| 99热这里只有精品一区| 国产精品久久久久久亚洲av鲁大| 亚洲欧美清纯卡通| 亚洲最大成人中文| 亚洲av免费在线观看| 国内精品久久久久久久电影| 99视频精品全部免费 在线| 最近中文字幕高清免费大全6 | 性色av乱码一区二区三区2| 高潮久久久久久久久久久不卡| 最新中文字幕久久久久| 日韩 亚洲 欧美在线| 国产视频一区二区在线看| 18禁在线播放成人免费| 午夜精品一区二区三区免费看| 国产一区二区在线观看日韩| 成年女人毛片免费观看观看9| 久久人人精品亚洲av| 热99re8久久精品国产| 此物有八面人人有两片| 麻豆av噜噜一区二区三区| 成人三级黄色视频| 亚洲av不卡在线观看| 国产精品影院久久| 88av欧美| av在线老鸭窝| 国产国拍精品亚洲av在线观看| 久久精品国产亚洲av天美| 搡女人真爽免费视频火全软件 | 国产精品乱码一区二三区的特点| 中文字幕熟女人妻在线| 一本综合久久免费| 欧美zozozo另类| 精品不卡国产一区二区三区| 免费看a级黄色片| 日本 欧美在线| 成年版毛片免费区| 两个人视频免费观看高清| 亚洲,欧美,日韩| 淫秽高清视频在线观看| 日韩大尺度精品在线看网址| 宅男免费午夜| 丰满人妻熟妇乱又伦精品不卡| avwww免费| 精品久久久久久成人av| 丁香欧美五月| 国产成人欧美在线观看| 国产精品久久久久久亚洲av鲁大| 国产美女午夜福利| 一进一出抽搐gif免费好疼| 久久午夜亚洲精品久久| 大型黄色视频在线免费观看| 国产69精品久久久久777片| 久久国产乱子免费精品| 精品久久久久久成人av| 免费电影在线观看免费观看| 国产在线精品亚洲第一网站| 色噜噜av男人的天堂激情| 亚洲第一欧美日韩一区二区三区| 人人妻人人澡欧美一区二区| 老司机午夜十八禁免费视频| 精品人妻偷拍中文字幕| 禁无遮挡网站| 九九在线视频观看精品| 毛片一级片免费看久久久久 | 99久久无色码亚洲精品果冻| 黄色配什么色好看| 亚洲欧美日韩卡通动漫| 天天一区二区日本电影三级| 观看美女的网站| 亚洲最大成人中文| 亚洲黑人精品在线| 国产黄片美女视频| 亚洲精华国产精华精| 免费无遮挡裸体视频| 亚洲国产精品成人综合色| 国产黄片美女视频| 一个人观看的视频www高清免费观看| 日韩精品中文字幕看吧| 国产精品久久久久久人妻精品电影| 成人高潮视频无遮挡免费网站| 日本a在线网址| 久久久久国产精品人妻aⅴ院| 欧美成人性av电影在线观看| 日韩国内少妇激情av| 欧美成人性av电影在线观看| 日韩精品中文字幕看吧| 国产精品久久久久久人妻精品电影| 白带黄色成豆腐渣| 精品久久久久久久末码| 欧美区成人在线视频| 免费观看的影片在线观看| 精品一区二区三区人妻视频| 成人av在线播放网站| 久久热精品热| 国产日本99.免费观看| 亚洲自拍偷在线| 极品教师在线视频| 美女免费视频网站| 亚洲av电影不卡..在线观看| 3wmmmm亚洲av在线观看| 别揉我奶头 嗯啊视频| 日韩欧美一区二区三区在线观看| 精品99又大又爽又粗少妇毛片 |