5
A REAL-TIME DEEP NETWORK FOR CROWD COUNTING Xiaowen Shi , Xin Li †‡ , Caili Wu †‡ , Shuchen Kong ? , Jing Yang , Liang He †‡ Shanghai Key Laboratory of Multidimensional Information Processing East China Normal University, Shanghai, China ? Videt Tech Ltd., Shanghai, China ABSTRACT Automatic analysis of highly crowded people has attracted extensive attention from computer vision research. Previous approaches for crowd counting have already achieved promis- ing performance across various benchmarks. However, to deal with the real situation, we hope the model run as fast as possible while keeping accuracy. In this paper, we propose a compact convolutional neural network for crowd counting which learns a more efficient model with a small number of parameters. With three parallel filters executing the convo- lutional operation on the input image simultaneously at the front of the network, our model could achieve nearly real- time speed and save more computing resources. Experiments on two benchmarks show that our proposed method not only takes a balance between performance and efficiency which is more suitable for actual scenes but also is superior to existing light-weight models in speed. Index TermsCrowd counting, compact convolutional neural network. 1. INTRODUCTION By analyzing and understanding the crowd behavior and con- gestion levels in detail, some preventable calamities such as the stampede could be alleviated, which make great sense for public security. A strong demand to develop a responsive and efficient crowd counting application to effectively control the harm of emergencies is increasing and brings a big challenge to this vision task. The existing methods to address crowd counting problem could be divided into two groups: count-oriented approaches and density-oriented approaches. Count-oriented approaches simply output the number of people by using a detector to detect objects in a sliding window that glides across the en- tire image. However, when the density of crowd is extremely dense, the spatial distributions are almost totally different in each image, which makes count-oriented approach invalid. In this way, spatial information is displayed colorred in the form ofthrough the density map to indicate the amount of people across the whole image. This density map provides more ac- curate and comprehensive information, which could be a cru- cial part of making correct decisions in highly varied crowded scenes. With recent development of the convolutional neural net- work (CNN), researchers employ CNN to accurately estimate the crowd count from images or videos [1–5]. However, it is always challenging to deal with scale variations on static im- ages, especially in diversified scenes such as different cam- era perspectives and irregular crowd clusters. For this rea- son, many previous works investigate the multi-scale archi- tectures [1, 2, 6, 7] as the backbone to deal with this problem. Although they outperform than other kinds of method yet, this kind of complex models with a large number of parameters will probably cause time-consuming and sub-optimal prob- lem, which would be inappropriate for applications need fast response. To sum up, it is still far from the balance of accu- racy and efficiency required in actual scene. In this paper, we propose the compact convolutional neural network (C-CNN) to simplify multi-branch of CNN model. It involves a small number of parameters and achieves satisfying performance. The network utilizes three filters with different sizes of local receptive field in one layer. The generative feature maps are merged directly after re- ceptive fields, and then fed into a CNN structure to fit a density map.Compared with existing methods, the proposed model achieves substantially enhanced performance with faster speed and the maintain of accuracy. The contributions of this work are as follows: • Our network is simple enough to be trained efficiently when compare with those multi CNN frameworks which need pre-trained model in each branch. In addi- tion, it requires less computing resources and is more practical. • We obtain an optimal balance between the efficiency of the model and the accuracy of the estimated count, ensuring that our model can achieve the accurate results effectively. 2. RELATED WORK Traditional approaches. The early researches [8] adopted a detection-style framework to carry out the function of count- arXiv:2002.06515v1 [cs.CV] 16 Feb 2020

, Xin Li , Caili Wu , Shuchen Kong , Jing Yang , Liang He · Xiaowen Shiy, Xin Li yz, Caili Wu , Shuchen Kong?, Jing Yangy, Liang Heyz zShanghai Key Laboratory of Multidimensional

  • Upload
    others

  • View
    13

  • Download
    0

Embed Size (px)

Citation preview

Page 1: , Xin Li , Caili Wu , Shuchen Kong , Jing Yang , Liang He · Xiaowen Shiy, Xin Li yz, Caili Wu , Shuchen Kong?, Jing Yangy, Liang Heyz zShanghai Key Laboratory of Multidimensional

A REAL-TIME DEEP NETWORK FOR CROWD COUNTING

Xiaowen Shi†, Xin Li†‡, Caili Wu†‡, Shuchen Kong?, Jing Yang†, Liang He†‡

‡Shanghai Key Laboratory of Multidimensional Information Processing†East China Normal University, Shanghai, China

?Videt Tech Ltd., Shanghai, China

ABSTRACT

Automatic analysis of highly crowded people has attractedextensive attention from computer vision research. Previousapproaches for crowd counting have already achieved promis-ing performance across various benchmarks. However, todeal with the real situation, we hope the model run as fastas possible while keeping accuracy. In this paper, we proposea compact convolutional neural network for crowd countingwhich learns a more efficient model with a small number ofparameters. With three parallel filters executing the convo-lutional operation on the input image simultaneously at thefront of the network, our model could achieve nearly real-time speed and save more computing resources. Experimentson two benchmarks show that our proposed method not onlytakes a balance between performance and efficiency which ismore suitable for actual scenes but also is superior to existinglight-weight models in speed.

Index Terms— Crowd counting, compact convolutionalneural network.

1. INTRODUCTION

By analyzing and understanding the crowd behavior and con-gestion levels in detail, some preventable calamities such asthe stampede could be alleviated, which make great sense forpublic security. A strong demand to develop a responsive andefficient crowd counting application to effectively control theharm of emergencies is increasing and brings a big challengeto this vision task.

The existing methods to address crowd counting problemcould be divided into two groups: count-oriented approachesand density-oriented approaches. Count-oriented approachessimply output the number of people by using a detector todetect objects in a sliding window that glides across the en-tire image. However, when the density of crowd is extremelydense, the spatial distributions are almost totally different ineach image, which makes count-oriented approach invalid. Inthis way, spatial information is displayed colorred in the formofthrough the density map to indicate the amount of peopleacross the whole image. This density map provides more ac-curate and comprehensive information, which could be a cru-

cial part of making correct decisions in highly varied crowdedscenes.

With recent development of the convolutional neural net-work (CNN), researchers employ CNN to accurately estimatethe crowd count from images or videos [1–5]. However, it isalways challenging to deal with scale variations on static im-ages, especially in diversified scenes such as different cam-era perspectives and irregular crowd clusters. For this rea-son, many previous works investigate the multi-scale archi-tectures [1, 2, 6, 7] as the backbone to deal with this problem.Although they outperform than other kinds of method yet, thiskind of complex models with a large number of parameterswill probably cause time-consuming and sub-optimal prob-lem, which would be inappropriate for applications need fastresponse. To sum up, it is still far from the balance of accu-racy and efficiency required in actual scene.

In this paper, we propose the compact convolutionalneural network (C-CNN) to simplify multi-branch of CNNmodel. It involves a small number of parameters and achievessatisfying performance. The network utilizes three filterswith different sizes of local receptive field in one layer.The generative feature maps are merged directly after re-ceptive fields, and then fed into a CNN structure to fit adensity map.Compared with existing methods, the proposedmodel achieves substantially enhanced performance withfaster speed and the maintain of accuracy.

The contributions of this work are as follows:

• Our network is simple enough to be trained efficientlywhen compare with those multi CNN frameworkswhich need pre-trained model in each branch. In addi-tion, it requires less computing resources and is morepractical.

• We obtain an optimal balance between the efficiencyof the model and the accuracy of the estimated count,ensuring that our model can achieve the accurate resultseffectively.

2. RELATED WORK

Traditional approaches. The early researches [8] adopted adetection-style framework to carry out the function of count-

arX

iv:2

002.

0651

5v1

[cs

.CV

] 1

6 Fe

b 20

20

Page 2: , Xin Li , Caili Wu , Shuchen Kong , Jing Yang , Liang He · Xiaowen Shiy, Xin Li yz, Caili Wu , Shuchen Kong?, Jing Yangy, Liang Heyz zShanghai Key Laboratory of Multidimensional

ing. These methods detected the existence of a pedestrianin a sliding window by training a classifier using featuresextracted from a complete pedestrian. But it is difficult tocount the exact number of people if most of the target ob-jects are seriously obscured in highly congested scenes. Inthis case, researchers began to use specific body parts featuresto construct boosted classifiers [9]. Although the detection-based approaches have been improved though this modifica-tion, the perform is still poor in extremely dense situation, soresearchers tried to design regression-based approaches to di-rectly map the features extracted from image patches to scalarvalues [10, 11].

Nevertheless, regression-based methods can not perceivecrowd distributions as they ignored important spatial informa-tion and regressed on the global count. Density estimation-based approaches are therefore developed with the abilityto conduct pixel-wise regressions. Linear mapping [12] andnon-linear mapping [13] methods were utilized for densitycalculation successively.

CNN-based methods. With the breakthrough of deep learn-ing in computer vision [14], some researchers tried to useconvolutional neural network as feature extractor for crowdcounting task [1,7,15]. They adopted multiple CNN brancheswith different receptive fields to enable multi-scale adapta-tion and then combined the output feature maps of differentlevel of a congested scene and mapped them to a density map.These methods exactly obtained excellent performance on thehighly congested scene, but they need to pre-train each single-network for global optimization. Also, the branch structurefor learning different features for each column is inefficient,the redundant parameters have a negative impact on the fi-nal performance. Moreover, this kind of model is inactive inreal-world because of the low speed and high latency in in-ference. As a remedy, single-branch counting networks withscale adaptations were proposed. Cao et al. [16] computedhigh-quality maps with a new encoder-decoder network, aswell as a SIM local pattern consistent loss. However, it stillsuffers from a large number of parameters.

Unlike approaches mentioned above, the work in this pa-per is specifically aimed at reducing the number of parame-ters of the network by designing a sparse network structure.Specifically, we use three stacked filers of different size anddirectly target a merged feature map at once. In this way wecan utilize sparsity at the filter level to optimize parallel com-puting and increase network adaptability to scale, making itspend less time on training and optimization.

3. PROPOSED APPROACH

3.1. Compact Convolutional Neural Network

In a typical multi-scale CNN architecture, features extractedfrom multi-column CNN with different size of receptivefields. However, we perform multiple convolutions in parallel

9�

10�

9�

7�

14�

7�

5�

16�

5�

3�

3�

40�

3�

3�

60�

Max Pooling�

3�

3�

40�

3�

3�

20�

3�3�

10�

= Max Pooling�

Fig. 1: Overview of the proposed C-CNN architecture. Thenetwork incorporates three parallel filters of different sizein the color of red, green and blue respectively, which aremerged to estimate crowd density estimation.

on the input image and combine all the output results intoa very deep feature map in order to avoid parameters of ourmodel increase explosively.

The overall architecture of our framework is illustrated inFigure 1. The network could be divided into two components:the parallel convolution layer with different kernels and theconvolution or pooling layers that followed. In the front part,the red layer in the figure 1 is designed to pay more attentionto large receptive fields. The green one is on behalf of densecrowds and the blue one stands for highly congested crowdedscenes. All of them are followed by a 2 × 2 max-pollinglayer. After the extraction process of various receptive fields,the feature maps are merged as feature fusion for follow-uplayers to perform down-sampling. We find that using onlyone layer of convolution is enough to extract different spatialfeatures and could ameliorate the efficiency of feature extrac-tion from multiple branches. This is why the network is fasterand accurate. The latter part consists of 6 convolutional layersspecifically. Note that the third layer and the fourth layer arefollowed by a max-pooling layer with another 2 × 2 kernelsize. The last convolution layer uses 1 × 1 filter to aggregatethe feature maps into a density map.

Compared with those multi-column CNNs, our methodhas many improvements. Firstly, we find that merging featuremaps after first receptive fields in the head of the networkoutperforms connecting continued convolutional operations.Through our analysis, we consider that only using one layerof convolution could extract more comprehensive details ofimages, while the whole convolutional neural network doeswell on capturing local features but it would disrupt the spa-tial information. Another advantage is that comparing withthe multi-column architecture which is always puzzled by theredundancy and repetition of the number of filters in each col-umn, our approach can be seen as discarding the extra convo-lution operations. Thirdly, our approach is proved to makea reasonable trade-off between model performance and thenumber of network parameters. The experiments validate theeffectiveness of the proposed structure.

Page 3: , Xin Li , Caili Wu , Shuchen Kong , Jing Yang , Liang He · Xiaowen Shiy, Xin Li yz, Caili Wu , Shuchen Kong?, Jing Yangy, Liang Heyz zShanghai Key Laboratory of Multidimensional

3.2. Implementation Details

3.2.1. Ground truth generation

We use the geometry-adaptive kernels to generate a groundtruth density map of the highly congested scenes. Adaptinga Gaussian kernel each head annotation becomes vague, sothat we generate the ground truth density maps with the spa-tial distribution information across the whole image. Thismethod alleviates the difficulty of the regression becausewe could get more accurate and comprehensive informationrather than predict the exact point of head annotation directly.The geometry-adaptive kernels are defined as

F (x) =

N∑i=1

δ(x− xi)×Gσi(x),

where σi = βd̄i, xi stand for a target object in the groundtruth δ, and convert δ(x− xi) into a density map, we con-volve this with a Gaussian kernel with a standard deviation ofσi. Here d̄i means the average distance of k nearest neighborsof target object xi. In this experiment, we create density mapswith σi = 15.

3.2.2. Training details

The feature maps output from our model are mapped to thedensity maps adopting filters of size 1 × 1, then we use Eu-clidean distance to measure the difference between the outputdensity map and the corresponding ground truth. Here wedefine the loss function as follows,

L(Θ) =1

N

N∑i=1

||f(Xi,Θ)− Fi||2,

Where Fi is the ground truth density map of image Xi, thef(Xi,Θ) is an estimated density map which is parameterizedwith Θ for the sample Xi. During training, we set batch sizeto 8 and use Adam with learning rate of 10−5.

4. EXPERIMENTS

4.1. Results and Comparison

We conduct a comprehensive study using the ShanghaiTechdataset [7]and The WorldExpo’10 dataset [17]. We denoteour approach as C-CNN in the following comparisons anduse the MAE and MSE as evaluation metric.

ShanghaiTech dataset [7] is with 1198 images and 330,165annotated people. The dataset consists of Part A and Part B.Part A includes 482 crowd images with 300 training imagesand 182 testing images, while Part B contains 716 imageswhich divided into a training set with 400 images and test-ing set with 316 images. First, we evaluate and compare ourmethod with other four lightweight networks and the results

Table 1: Comparison on ShanghaiTech dataset.

MethodPart A Part B Parameter

MAE MSE MAE MSE sizeCMTL [18] 101.3 152.4 20.0 31.1 2.36MZhang et al. [17] 181.8 277.7 32.0 49.8 0.62MMCNN [7] 110.2 173.2 26.4 41.3 0.15MTDF-CNN [19] 97.5 145.1 20.7 32.8 0.13MC-CNN 88.1 141.7 14.9 22.1 0.07MACSCP [20] 75.7 102.7 17.2 27.4 5.10MSwitching CNN [1] 90.4 135.0 21.6 33.4 15.30MCSRNet [21] 68.3 115.0 10.6 16.0 16.26MSaCNN [22] 86.8 139.2 16.2 25.8 24.06MCP-CNN [2] 73.6 106.4 20.1 30.1 68.40M

Table 2: Comparison of C-CNN with both lightweight andlarge networks on WorldExpo’10 dataset.

Method S1 S2 S3 S4 S5 Avg. ParamsZhang et al. [17] 9.8 14.1 14.3 22.2 3.7 12.9 0.62MMCNN [7] 3.4 20.6 12.9 13.0 8.1 11.6 0.15MTDF-CNN [19] 2.7 23.4 10.7 17.6 3.3 11.5 0.13MC-CNN(ours) 3.8 20.5 8.8 8.8 7.7 9.9 0.07MCSRNet [21] 2.9 11.5 8.6 16.6 3.4 8.6 16.26MSaCNN [22] 2.6 13.5 10.6 12.5 3.3 8.5 24.06MCP-CNN [2] 2.9 14.7 10.5 10.4 5.8 8.86 68.40M

are shown in the upper part of Table 1. It displays that C-CNNwith simple architecture achieves the lowest MAE in Part Aand both of the lowest MAE and MSE in Part B. Note thatthe parameter size of our model is still the smallest one. Wealso compare with some large network and the results in thebottom of Table 1. Although the deeper models achieve betterperformance, their parameter size is around 200 times morethan ours. Some qualitative results are presented in Figure 2.

WorldExpo’10 dataset [17] contains 1,132 annotated imagesthat were captured in the 2010 Shanghai WorldExpo by 108surveillance cameras. The dataset is divided into 5 differentscenes, marked as S1 to S5 in Table 2. Our C-CNN deliversthe best results in scene 3 and scene 4 when comparing itwith other light models, displayed in the top part of the table.Besides, it also achieves the best accuracy on average withthe smallest parameter size. Compare with large networks,our approach outperforms them in scene 4 with even tens oftimes smaller parameters. This representation reinforces thefact that C-CNN is light enough without sacrificing too muchaccuracy.

4.2. Ablation Study

In this part, we perform an ablation study to analyze the pro-posed framework on ShanghaiTech Part A dataset.

Network architecture. To evaluate the effect of structuralvariations of the three filters of different sizes, we separately

Page 4: , Xin Li , Caili Wu , Shuchen Kong , Jing Yang , Liang He · Xiaowen Shiy, Xin Li yz, Caili Wu , Shuchen Kong?, Jing Yangy, Liang Heyz zShanghai Key Laboratory of Multidimensional

T:414.59 red: 404.45

T:796.57 red: 786.90

GT:399.90 red: 392.22

GT 364.10 red: 349.10

(a) ShanghaiTech Part A

T:278.22 red: 271.93

T:417.58 red: 414.66

T:150.11 red: 149.93

GT:111.05 red: 110.85

(b) ShanghaiTech Part B

Fig. 2: Qualitative results on the benchmarks.

Fig. 3: the effect of varying network architecture, a. 5×5filter only; b. 7×7 filter only; c. 9×9 filter only; d. learnedweight without last pooling layer; e. proposed approach.

train C-CNN with choice of filters with the size of 5 × 5, 7× 7, 9 × 9 respectively and all three filters contained concur-rently. Figure 3 shows the comparison result. We observe thatC-CNN with the filter of size 5× 5 only outperforms with thefilter of 7 × 7 or the filter of 9 × 9. It is easy to ascribe tofilter of size 5 × 5 capture crowds at lower scales within thescene and be more advantageous to extract the characteristicsof highly congested crowded scenes. With the formation of allthree filters contained, the result is 88.08 of MAE and 141.72of MSE, which is the lowest. It shows that the scale and per-spective variations could be adapted better with the structureof three independent filters of respective sizes.

Effect of the pooling operation. From Figure 3 we can ob-serve that without the last pooling operation, we obtain theMAE of 98.9 and MSE of 160.4, which is inferior to the com-plete model. This evidence tells the fact that the last poolinglayer plays an indispensable role in the whole model, and thecurrent architecture has reached its peak of balance betweenoperational speed and prediction accuracy. This is becausethe pooling layer provides the characteristic of scale invari-

Table 3: The speed of different methods.Method CMTL [18] MCNN [7] C-CNNFPS 8.37 64.52 104.16

ance to local translation which can help us pay more attentionto the existence of the feature other than the exact location ofit, especially in the field of crowd counting.

4.3. Speed Comparison

In this section, we compare C-CNN with the other two crowdcounting methods MCNN and CMTL. The main reason of us-ing MCNN in [7] and CMTL in [18] is the relatively smallnumber of parameters provided by the whole network withthe lower MAE and MSE displayed in Table 1. Here we useFPS(Frame per second), which is the most commonly usedevaluation metrics for measuring the speed of models to ac-cess our model fairly. All three methods are tested on Shang-haiTech PartA and PartB, and the figures we reported are cal-culated running through 182 test images on the ShanghaiTechPart A dataset and 316 test images on the ShanghaiTech PartB dataset at their original resolution (768×1024). Further-more, these experiments were all conducted under the samecondition with a server using the GPU (GeForce GTX 1080)and the CPU (Intel i5-8500 @ 3.00 GHz × 6). The over-all speed comparison with the other state-of-the-art models isdemonstrated in Table 3.

As can be seen from Table 3, our method gain the highestscore at an average speed of 104.16 fps, which is much higherthan other methods. with the cost time of 2.14s on the entiretesting set from ShanghaiTech Part A and 4.39s on Part B, thespeed on prediction phase even achieve more than 10 timesfaster than that of other advanced models. All these confirmthat our work is valuable because only with high speed canreal-time processing be realized, which is extremely impor-tant for some application scenarios.

5. CONCLUSIONS

In this paper, we present a compact CNN for crowd count-ing to deal with the lack of real-time performance of existingmethods. By removing the redundant and recurrent convolu-tional layers and designing a superior local sparse structure,the parameter size is significantly reduced. Specifically, weusing a multiple juxtaposed convolution structure where fea-ture maps extracted from three parallel convolutional layerswith different size of receptive fields are directly fused. Com-pared with the baseline approaches, the proposed model ob-tains an improvement significantly.

Page 5: , Xin Li , Caili Wu , Shuchen Kong , Jing Yang , Liang He · Xiaowen Shiy, Xin Li yz, Caili Wu , Shuchen Kong?, Jing Yangy, Liang Heyz zShanghai Key Laboratory of Multidimensional

6. REFERENCES

[1] D. B. Sam, S. Surya, and R. V. Babu, “Switching convo-lutional neural network for crowd counting,” in CVPR,2017, pp. 4031–4039.

[2] V. A. Sindagi and V. M. Patel, “Generating high-qualitycrowd density maps using contextual pyramid cnns,” inICCV, 2017, pp. 1861–1870.

[3] E. Walach and L. Wolf, “Learning to count with cnnboosting,” in ECCV, 2016, pp. 660–676.

[4] X. Wu, Y. Zheng, H. Ye, W. Hu, J. Yang, and L. He,“Adaptive scenario discovery for crowd counting,” inICASSP, 2019, pp. 2382–2386.

[5] X. Wu, B. Xu, Y. Zheng, H. Ye, J. Yang, and L. He, “Fastvideo crowd counting with a temporal aware network,”arXiv:1907.02198, 2019.

[6] L. Boominathan, S. S. Kruthiventi, and R. V. Babu,“Crowdnet: A deep convolutional network for densecrowd counting,” in ACM MM, 2016, pp. 640–644.

[7] Y. Zhang, D. Zhou, S. Chen, S. Gao, and Y. Ma, “Single-image crowd counting via multi-column convolutionalneural network,” in CVPR, 2016, pp. 589–597.

[8] P. Dollar, C. Wojek, B. Schiele, and P. Perona, “Pedes-trian detection: An evaluation of the state of the art,”IEEE Transaction on Pattern Analysis and Machine In-telligence, vol. 34, no. 4, pp. 743–761, 2012.

[9] B. Wu and R. Nevatia, “Detection and tracking of multi-ple, partially occluded humans by bayesian combinationof edgelet based part detectors,” International Journalof Computer Vision, vol. 75, no. 2, pp. 247–266, 2007.

[10] A. B. Chan and N. Vasconcelos, “Bayesian poisson re-gression for crowd counting,” in ICCV, 2009, pp. 545–551.

[11] H. Idrees, I. Saleemi, C. Seibert, and M. Shah, “Multi-source multi-scale counting in extremely dense crowdimages,” in CVPR, 2013, pp. 2547–2554.

[12] V. Lempitsky and A. Zisserman, “Learning to count ob-jects in images,” in NeurIPS, 2010, pp. 1324–1332.

[13] V.-Q. Pham, T. Kozakaya, O. Yamaguchi, and R. Okada,“Count forest: Co-voting uncertain number of targetsusing random forest for crowd density estimation,” inICCV, 2015, pp. 3253–3261.

[14] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Ima-genet classification with deep convolutional neural net-works,” in NeurIPS, 2012, pp. 1097–1105.

[15] D. Babu Sam, N. N. Sajjan, R. Venkatesh Babu, andM. Srinivasan, “Divide and grow: Capturing huge diver-sity in crowd images with incrementally growing cnn,”in CVPR, 2018, pp. 3618–3626.

[16] X. Cao, Z. Wang, Y. Zhao, and F. Su, “Scale aggregationnetwork for accurate and efficient crowd counting,” inECCV, 2018, pp. 734–750.

[17] C. Zhang, H. Li, X. Wang, and X. Yang, “Cross-scene crowd counting via deep convolutional neural net-works,” in CVPR, 2015, pp. 833–841.

[18] V. A. Sindagi and V. M. Patel, “Cnn-based cascadedmulti-task learning of high-level prior and density esti-mation for crowd counting,” in AVSS, 2017, pp. 1–6.

[19] D. B. Sam and R. V. Babu, “Top-down feedback forcrowd counting convolutional neural network,” in AAAI,2018.

[20] Z. Shen, Y. Xu, B. Ni, M. Wang, J. Hu, and X. Yang,“Crowd counting via adversarial cross-scale consistencypursuit,” in CVPR, 2018, pp. 5245–5254.

[21] Y. Li, X. Zhang, and D. Chen, “Csrnet: Dilated con-volutional neural networks for understanding the highlycongested scenes,” in CVPR, 2018, pp. 1091–1100.

[22] L. Zhang, M. Shi, and Q. Chen, “Crowd counting viascale-adaptive convolutional neural network,” in WACV,2018, pp. 1113–1121.