日韩性视频-久久久蜜桃-www中文字幕-在线中文字幕av-亚洲欧美一区二区三区四区-撸久久-香蕉视频一区-久久无码精品丰满人妻-国产高潮av-激情福利社-日韩av网址大全-国产精品久久999-日本五十路在线-性欧美在线-久久99精品波多结衣一区-男女午夜免费视频-黑人极品ⅴideos精品欧美棵-人人妻人人澡人人爽精品欧美一区-日韩一区在线看-欧美a级在线免费观看

歡迎訪問 生活随笔!

生活随笔

當前位置: 首頁 > 编程资源 > 编程问答 >内容正文

编程问答

2019计算机视觉论文精选速递(2019/1/23-2018/1/28)

發布時間:2025/3/8 编程问答 37 豆豆
生活随笔 收集整理的這篇文章主要介紹了 2019计算机视觉论文精选速递(2019/1/23-2018/1/28) 小編覺得挺不錯的,現在分享給大家,幫大家做個參考.

作者:朱政
原文:CV arXiv Daily:計算機視覺論文每日精選(2019/1/23-2018/1/28)
如有興趣可以**點擊加入極市CV專業微信群**,獲取更多高質量干貨

本系列文章轉自計算機視覺牛人朱政大佬的微信公眾號(CV arxiv Daily),已經授權轉載,主要是為了幫大家篩選計算機視覺領域每天的arXiv中的論文,主要關注領域:目標檢測,圖像分割,單/多目標跟蹤,行為識別,人體姿態估計與跟蹤,行人重識別,GAN,模型搜索等。歡迎關注我,每日會定時轉發,努力學習起來~

2019/1/28

[1] Google的自監督表征學習文章
Revisiting Self-Supervised Visual Representation Learning
論文鏈接:https://arxiv.org/abs/1901.09005
代碼地址:https://github.com/google/revisiting-self-supervised
摘要: Unsupervised visual representation learning remains a largely unsolved problem in computer vision research. Among a big body of recently proposed approaches for unsupervised learning of visual representations, a class of self-supervised techniques achieves superior performance on many challenging benchmarks. A large number of the pretext tasks for self-supervised learning have been studied, but other important aspects, such as the choice of convolutional neural networks (CNN), has not received equal attention. Therefore, we revisit numerous previously proposed self-supervised models, conduct a thorough large scale study and, as a result, uncover multiple crucial insights. We challenge a number of common practices in selfsupervised visual representation learning and observe that standard recipes for CNN design do not always translate to self-supervised representation learning. As part of our study, we drastically boost the performance of previously proposed techniques and outperform previously published state-of-the-art results by a large margin.


[2] ICLR 2019 GAN文章
Diversity-Sensitive Conditional Generative Adversarial Networks
論文鏈接:https://arxiv.org/abs/1901.09024
摘要: We propose a simple yet highly effective method that addresses the mode-collapse problem in the Conditional Generative Adversarial Network (cGAN). Although conditional distributions are multi-modal (i.e., having many modes) in practice, most cGAN approaches tend to learn an overly simplified distribution where an input is always mapped to a single output regardless of variations in latent code. To address such issue, we propose to explicitly regularize the generator to produce diverse outputs depending on latent codes. The proposed regularization is simple, general, and can be easily integrated into most conditional GAN objectives. Additionally, explicit regularization on generator allows our method to control a balance between visual quality and diversity. We demonstrate the effectiveness of our method on three conditional generation tasks: image-to-image translation, image inpainting, and future video prediction. We show that simple addition of our regularization to existing models leads to surprisingly diverse generations, substantially outperforming the previous approaches for multi-modal conditional generation specifically designed in each individual task.


[3] 上交盧策吾老師的Q-learning for斗地主 文章
Combinational Q-Learning for Dou Di Zhu
論文鏈接:https://arxiv.org/abs/1901.08925
代碼地址:https://github.com/qq456cvb/doudizhu-C
摘要: Deep reinforcement learning (DRL) has gained a lot of attention in recent years, and has been proven to be able to play Atari games and Go at or above human levels. However, those games are assumed to have a small fixed number of actions and could be trained with a simple CNN network. In this paper, we study a special class of Asian popular card games called Dou Di Zhu, in which two adversarial groups of agents must consider numerous card combinations at each time step, leading to huge number of actions. We propose a novel method to handle combinatorial actions, which we call combinational Q-learning (CQL). We employ a two-stage network to reduce action space and also leverage order-invariant max-pooling operations to extract relationships between primitive actions. Results show that our method prevails over state-of-the art methods like naive Q-learning and A3C. We develop an easy-to-use card game environments and train all agents adversarially from sractch, with only knowledge of game rules and verify that our agents are comparative to humans. Our code to reproduce all reported results will be available online.


[4] WACV2019 3D點云 文章
Dense 3D Point Cloud Reconstruction Using a Deep Pyramid Network
論文鏈接:https://arxiv.org/abs/1901.08906
摘要: Reconstructing a high-resolution 3D model of an object is a challenging task in computer vision. Designing scalable and light-weight architectures is crucial while addressing this problem. Existing point-cloud based reconstruction approaches directly predict the entire point cloud in a single stage. Although this technique can handle low-resolution point clouds, it is not a viable solution for generating dense, high-resolution outputs. In this work, we introduce DensePCR, a deep pyramidal network for point cloud reconstruction that hierarchically predicts point clouds of increasing resolution. Towards this end, we propose an architecture that first predicts a low-resolution point cloud, and then hierarchically increases the resolution by aggregating local and global point features to deform a grid. Our method generates point clouds that are accurate, uniform and dense. Through extensive quantitative and qualitative evaluation on synthetic and real datasets, we demonstrate that DensePCR outperforms the existing state-of-the-art point cloud reconstruction works, while also providing a light-weight and scalable architecture for predicting high-resolution outputs.


[5] Multi-Target Multi-Camera Tracking 文章
Multiple Hypothesis Tracking Algorithm for Multi-Target Multi-Camera Tracking with Disjoint Views
論文鏈接:https://arxiv.org/abs/1901.08787
摘要: In this study, a multiple hypothesis tracking (MHT) algorithm for multi-target multi-camera tracking (MCT) with disjoint views is proposed. Our method forms track-hypothesis trees, and each branch of them represents a multi-camera track of a target that may move within a camera as well as move across cameras. Furthermore, multi-target tracking within a camera is performed simultaneously with the tree formation by manipulating a status of each track hypothesis. Each status represents three different stages of a multi-camera track: tracking, searching, and end-of-track. The tracking status means targets are tracked by a single camera tracker. In the searching status, the disappeared targets are examined if they reappear in other cameras. The end-of-track status does the target exited the camera network due to its lengthy invisibility. These three status assists MHT to form the track-hypothesis trees for multi-camera tracking. Furthermore, they present a gating technique for eliminating of unlikely observation-to-track association. In the experiments, they evaluate the proposed method using two datasets, DukeMTMC and NLPR-MCT, which demonstrates that the proposed method outperforms the state-of-the-art method in terms of improvement of the accuracy. In addition, they show that the proposed method can operate in real-time and online.


[6] One-Class CNN 文章
One-Class Convolutional Neural Network
論文鏈接:https://arxiv.org/abs/1901.08688
代碼地址:github.com/otkupjnoz/oc-cnn
摘要: We present a novel Convolutional Neural Network (CNN) based approach for one class classification. The idea is to use a zero centered Gaussian noise in the latent space as the pseudo-negative class and train the network using the cross-entropy loss to learn a good representation as well as the decision boundary for the given class. A key feature of the proposed approach is that any pre-trained CNN can be used as the base network for one class classification. The proposed One Class CNN (OC-CNN) is evaluated on the UMDAA-02 Face, Abnormality-1001, FounderType-200 datasets. These datasets are related to a variety of one class application problems such as user authentication, abnormality detection and novelty detection. Extensive experiments demonstrate that the proposed method achieves significant improvements over the recent state-of-the-art methods. The source code is available at : github.com/otkupjnoz/oc-cnn.


[7] In Defense of the Triplet Loss 文章
In Defense of the Triplet Loss for Visual Recognition
論文鏈接:https://arxiv.org/abs/1901.08616
摘要: We employ triplet loss as a space embedding regularizer to boost classification performance. Standard architectures, like ResNet and DesneNet, are extended to support both losses with minimal hyper-parameter tuning. This promotes generality while fine-tuning pretrained networks. Triplet loss is a powerful surrogate for recently proposed embedding regularizers. Yet, it is avoided for large batch-size requirement and high computational cost. Through our experiments, we re-assess these assumptions. During inference, our network supports both classification and embedding tasks without any computational overhead. Quantitative evaluation highlights how our approach compares favorably to the existing state of the art on multiple fine-grained recognition datasets. Further evaluation on an imbalanced video dataset achieves significant improvement (>7%). Beyond boosting efficiency, triplet loss brings retrieval and interpretability to classification models.

2019/1/26

SiamRPN系列文章總結

[0] SiamFC文章,對SINT(Siamese Instance Search for Tracking,in CVPR2016)改進,第一個提出用全卷積孿生網絡結構來解決tracking問題的paper,可以視為只有一個anchor的SiamRPN
論文題目:Fully-convolutional siamese networks for object tracking
論文地址:https://arxiv.org/abs/1606.09549
項目地址:https://www.robots.ox.ac.uk/~luca/siamese-fc.html
tf實現:https://github.com/torrvision/siamfc-tf
pytorch實現:https://github.com/rafellerc/Pytorch-SiamFC


[0.1] 后面的v2版本即CFNet,用cf操作代替了correlation操作。
論文題目:End-To-End Representation Learning for Correlation Filter Based Tracking
論文地址:http://openaccess.thecvf.com/content_cvpr_2017/html/Valmadre_End-To-End_Representation_Learning_CVPR_2017_paper.html
項目地址:http://www.robots.ox.ac.uk/~luca/cfnet.html
MatConvNet實現:https://github.com/bertinetto/cfnet
SiamFC之后有諸多的改進工作,例如


[0.2] StructSiam,在跟蹤中考慮Local structures
論文題目:Structured Siamese Network for Real-Time Visual Tracking
論文地址:http://openaccess.thecvf.com/content_ECCV_2018/papers/Yunhua_Zhang_Structured_Siamese_Network_ECCV_2018_paper.pdf


[0.3] SiamFC-tri,在Saimese跟蹤網絡中引入了Triplet Loss
論文題目:Triplet Loss in Siamese Network for Object Tracking
論文地址:http://openaccess.thecvf.com/content_ECCV_2018/papers/Xingping_Dong_Triplet_Loss_with_ECCV_2018_paper.pdf


[0.4] DSiam,動態Siamese網絡
論文題目:Learning Dynamic Siamese Network for Visual Object Tracking
論文地址:http://openaccess.thecvf.com/content_ICCV_2017/papers/Guo_Learning_Dynamic_Siamese_ICCV_2017_paper.pdf
代碼地址:https://github.com/tsingqguo/DSiam


[0.5] SA-Siam,Twofold Siamese網絡
論文題目:A Twofold Siamese Network for Real-Time Object Tracking
論文地址:http://openaccess.thecvf.com/content_cvpr_2018/papers/He_A_Twofold_Siamese_CVPR_2018_paper.pdf


[1] SiamRPN文章,將anchor應用在候選區域的每個位置,同時進行分類和回歸,one-shot local detection。
論文題目:High Performance Visual Tracking with Siamese Region Proposal Network
論文地址:http://openaccess.thecvf.com/content_cvpr_2018/papers/Li_High_Performance_Visual_CVPR_2018_paper.pdf
項目地址:http://bo-li.info/SiamRPN/


[2] DaSiamRPN, SiamRPN文章的follow-up,重點強調了訓練過程中樣本不均衡的問題,增加了正樣本的種類和有語義的負樣本。
論文題目:Distractor-aware Siamese Networks for Visual Object Tracking
論文地址:https://arxiv.org/abs/1808.06048
項目地址:http://bo-li.info/DaSiamRPN/
test code:https://github.com/foolwood/DaSiamRPN


[3] Cascaded SiamRPN,將若干RPN模塊cascade起來,同時利用了不同layer的feature。
論文題目:Siamese Cascaded Region Proposal Networks for Real-Time Visual Tracking
論文地址:https://arxiv.org/abs/1812.06148


[4] SiamMask,在SiamRPN的結構中增加了一個mask分支,同時進行tracking和video segmentation。
論文題目:Fast Online Object Tracking and Segmentation: A Unifying Approach
論文地址:https://arxiv.org/abs/1812.05050
項目地址:http://www.robots.ox.ac.uk/~qwang/SiamMask/


[5] SiamRPN++, SiamRPN文章的follow-up,讓現代網絡例如ResNet在tracking中work了,基本上所有數據集都是SOTA。
論文題目:SiamRPN++: Evolution of Siamese Visual Tracking with Very Deep Networks
論文地址:https://arxiv.org/abs/1812.11703
項目地址:http://bo-li.info/SiamRPN++/


[6] Deeper and Wider SiamRPN,將網絡加深加寬來提升性能,重點關注感受野和padding的影響。
論文題目:Deeper and Wider Siamese Networks for Real-Time Visual Tracking
論文地址:https://arxiv.org/abs/1901.01660
test code:https://gitlab.com/MSRA_NLPR/deeper_wider_siamese_trackers

2019/1/25


[1] Salient Object Detection文章
Deep Reasoning with Multi-scale Context for Salient Object Detection
論文鏈接:https://arxiv.org/abs/1901.08362


[2] 交通場景異常檢測綜述
Anomaly Detection in Road Traffic Using Visual Surveillance: A Survey
論文鏈接:https://arxiv.org/abs/1901.08292


[3] 3D目標檢測
3D Backbone Network for 3D Object Detection
論文鏈接:https://arxiv.org/abs/1901.08373


[4] 語義分割文章
Application of Decision Rules for Handling Class Imbalance in Semantic Segmentation
論文鏈接:https://arxiv.org/abs/1901.08394


[5] 目標檢測文章
Object Detection based on Region Decomposition and Assembly
論文鏈接:https://arxiv.org/abs/1901.08225


[6] 牛津的圖卷積網絡文章
Hypergraph Convolution and Hypergraph Attention
論文鏈接:https://arxiv.org/abs/1901.08150

2019/1/24

[1] 京東PoseTrack2018亞軍方案的技術報告
A Top-down Approach to Articulated Human Pose Estimation and Tracking
論文鏈接:https://arxiv.org/abs/1901.07680


[2] 投稿TNNLS網絡壓縮文章
Towards Compact ConvNets via Structure-Sparsity Regularized Filter Pruning
論文鏈接:https://arxiv.org/abs/1901.07827
代碼:https://github.com/ShaohuiLin/SSR


[3] 港中文&商湯 DeepFashion數據集
DeepFashion2: A Versatile Benchmark for Detection, Pose Estimation, Segmentation and Re-Identification of Clothing Images
論文鏈接:https://arxiv.org/abs/1901.07973
代碼:https://github.com/switchablenorms/DeepFashion2

[4]目標檢測文章
Bottom-up Object Detection by Grouping Extreme and Center Points
論文鏈接:https://arxiv.org/abs/1901.08043
代碼:https://github.com/xingyizhou/ExtremeNet

2019/1/23

[1] 商湯 COCO2018 檢測任務冠軍方案文章
Winning entry of COCO 2018 Challenge (object detection task) Hybrid Task Cascade for Instance Segmentation
https://arxiv.org/abs/1901.07518


[2] 小米用NAS做超分的技術報告
Fast, Accurate and Lightweight Super-Resolution with Neural Architecture Search
https://arxiv.org/abs/1901.07261


[3] 目標檢測文章
Consistent Optimization for Single-Shot Object Detection
https://arxiv.org/abs/1901.06563


[4] 商湯的不均衡樣本分類文章
Dynamic Curriculum Learning for Imbalanced Data Classification
https://arxiv.org/abs/1901.06783


[5] 人臉檢測文章
Improved Selective Refinement Network for Face Detection
https://arxiv.org/abs/1901.06651


[6] 曠視的零售商品數據集
RPC: A Large-Scale Retail Product Checkout Dataset
https://arxiv.org/abs/1901.07249


[7] 人體屬性識別綜述
Pedestrian Attribute Recognition: A Survey
https://arxiv.org/abs/1901.07474
項目地址:https://sites.google.com/view/ahu-pedestrianattributes/


推薦文章

  • 目標檢測領域還有什么可做的?19 個方向給你建議
  • ECCV 2018 | CornerNet:目標檢測算法新思路

總結

以上是生活随笔為你收集整理的2019计算机视觉论文精选速递(2019/1/23-2018/1/28)的全部內容,希望文章能夠幫你解決所遇到的問題。

如果覺得生活随笔網站內容還不錯,歡迎將生活随笔推薦給好友。