Unpack `.tex` source.
Browse files- WireSegHR-tex.tar.gz +3 -0
- paper-tex/cvpr.sty +645 -0
- paper-tex/egbib.bib +592 -0
- paper-tex/figure_tex/additional_visualizations.tex +8 -0
- paper-tex/figure_tex/annotation.tex +11 -0
- paper-tex/figure_tex/comparison.tex +27 -0
- paper-tex/figure_tex/failure.tex +11 -0
- paper-tex/figure_tex/inpaint_compare.tex +12 -0
- paper-tex/figure_tex/motivation.tex +11 -0
- paper-tex/figure_tex/new_failure_cases.tex +9 -0
- paper-tex/figure_tex/new_panorama.tex +8 -0
- paper-tex/figure_tex/overpredict.tex +14 -0
- paper-tex/figure_tex/panorama.tex +9 -0
- paper-tex/figure_tex/pipeline.tex +9 -0
- paper-tex/figure_tex/pixel6.tex +10 -0
- paper-tex/figure_tex/teaser.tex +16 -0
- paper-tex/figures/additional_visualizations.pdf +3 -0
- paper-tex/figures/annotation.pdf +3 -0
- paper-tex/figures/inpainting_result.pdf +3 -0
- paper-tex/figures/new_failures.pdf +3 -0
- paper-tex/figures/new_panorama.pdf +3 -0
- paper-tex/figures/panorama.pdf +3 -0
- paper-tex/figures/pixel6.pdf +3 -0
- paper-tex/figures/qualitative.pdf +3 -0
- paper-tex/figures/teaser2.pdf +3 -0
- paper-tex/figures/wire-ablation.pdf +3 -0
- paper-tex/figures/wire-pipeline.png +3 -0
- paper-tex/figures/wire-pixel6.pdf +3 -0
- paper-tex/figures/wire-qualitative.pdf +3 -0
- paper-tex/figures/wireishard.pdf +3 -0
- paper-tex/ieee_fullname.bst +1135 -0
- paper-tex/ms.bbl +337 -0
- paper-tex/ms.tex +100 -0
- paper-tex/sections/abstract.tex +23 -0
- paper-tex/sections/conclusion.tex +8 -0
- paper-tex/sections/dataset.tex +32 -0
- paper-tex/sections/discussion.tex +43 -0
- paper-tex/sections/introduction.tex +86 -0
- paper-tex/sections/method.tex +98 -0
- paper-tex/sections/method_yq.tex +111 -0
- paper-tex/sections/related_work.tex +36 -0
- paper-tex/sections/results.tex +156 -0
- paper-tex/supplement.bbl +24 -0
- paper-tex/supplement.tex +184 -0
- paper-tex/tables/component.tex +25 -0
- paper-tex/tables/inpaint.tex +30 -0
- paper-tex/tables/logit.tex +18 -0
- paper-tex/tables/results.tex +53 -0
- paper-tex/tables/stats.tex +22 -0
- paper-tex/tables/thresholds.tex +22 -0
WireSegHR-tex.tar.gz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0a17096a2eaad07f51345426465697fcf0ee1a0c5b54aa6742b4ac23406f6bc4
|
| 3 |
+
size 33690376
|
paper-tex/cvpr.sty
ADDED
|
@@ -0,0 +1,645 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
% ---------------------------------------------------------------
|
| 2 |
+
%
|
| 3 |
+
% The last version of CVPR/ICCV LaTeX template had been developed
|
| 4 |
+
% by [email protected] and [email protected] about 15 years ago.
|
| 5 |
+
% That version suffered from several issues:
|
| 6 |
+
% 1. Authors needed several individual files: cvpr.sty,
|
| 7 |
+
% cvpr_eso.sty, eso-pic.sty.
|
| 8 |
+
% 2. For CVPR/ICCV rebuttals, another version of cvpr.sty was
|
| 9 |
+
% required.
|
| 10 |
+
% 3. Several warnings arose due to deprecated options.
|
| 11 |
+
%
|
| 12 |
+
% More recently, Ming-Ming Cheng ([email protected]) created
|
| 13 |
+
% a single style file that helps to unify review, rebuttal, and
|
| 14 |
+
% final versions with a class.
|
| 15 |
+
%
|
| 16 |
+
% This more recent style has been further modernized for CVPR 2022
|
| 17 |
+
% by Stefan Roth ([email protected]).
|
| 18 |
+
%
|
| 19 |
+
% Acknowledgements: This file is built on the template by
|
| 20 |
+
% Ming-Ming Cheng (https://github.com/MCG-NKU/CVPR_Template).
|
| 21 |
+
% ---------------------------------------------------------------
|
| 22 |
+
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
|
| 26 |
+
% ---------------------------------------------------------------
|
| 27 |
+
%
|
| 28 |
+
% $Id: cvpr.sty,v 1.3 2005/10/24 19:56:15 awf Exp $
|
| 29 |
+
% by [email protected] some mods by [email protected]
|
| 30 |
+
%
|
| 31 |
+
% ---------------------------------------------------------------
|
| 32 |
+
%
|
| 33 |
+
% no guarantee is given that the format corresponds perfectly to
|
| 34 |
+
% IEEE 8.5" x 11" Proceedings, but most features should be ok.
|
| 35 |
+
%
|
| 36 |
+
% ---------------------------------------------------------------
|
| 37 |
+
% with LaTeX2e:
|
| 38 |
+
% =============
|
| 39 |
+
%
|
| 40 |
+
% use as
|
| 41 |
+
% \documentclass[times,10pt,twocolumn]{article}
|
| 42 |
+
% \usepackage[options]{cvpr}
|
| 43 |
+
% \usepackage{times}
|
| 44 |
+
%
|
| 45 |
+
% "options" should be replaced by
|
| 46 |
+
% * "review" for submitting a paper for review,
|
| 47 |
+
% * "final" for the camera ready, and
|
| 48 |
+
% * "rebuttal" for the author rebuttal.
|
| 49 |
+
%
|
| 50 |
+
% specify references as
|
| 51 |
+
% {\small
|
| 52 |
+
% \bibliographystyle{ieee}
|
| 53 |
+
% \bibliography{...your files...}
|
| 54 |
+
% }
|
| 55 |
+
% ---------------------------------------------------------------
|
| 56 |
+
|
| 57 |
+
|
| 58 |
+
|
| 59 |
+
% ---------------------------------------------------------------
|
| 60 |
+
%
|
| 61 |
+
%\usepackage{eso-pic}
|
| 62 |
+
%
|
| 63 |
+
%%
|
| 64 |
+
%% This is file `eso-pic.sty',
|
| 65 |
+
%% generated with the docstrip utility.
|
| 66 |
+
%%
|
| 67 |
+
%% The original source files were:
|
| 68 |
+
%%
|
| 69 |
+
%% eso-pic.dtx (with options: `package')
|
| 70 |
+
%%
|
| 71 |
+
%% This is a generated file.
|
| 72 |
+
%%
|
| 73 |
+
%% Copyright (C) 1998-2002 by Rolf Niepraschk <[email protected]>
|
| 74 |
+
%%
|
| 75 |
+
%% This file may be distributed and/or modified under the conditions of
|
| 76 |
+
%% the LaTeX Project Public License, either version 1.2 of this license
|
| 77 |
+
%% or (at your option) any later version. The latest version of this
|
| 78 |
+
%% license is in:
|
| 79 |
+
%%
|
| 80 |
+
%% http://www.latex-project.org/lppl.txt
|
| 81 |
+
%%
|
| 82 |
+
%% and version 1.2 or later is part of all distributions of LaTeX version
|
| 83 |
+
%% 1999/12/01 or later.
|
| 84 |
+
%%
|
| 85 |
+
|
| 86 |
+
|
| 87 |
+
%
|
| 88 |
+
\NeedsTeXFormat{LaTeX2e}[1999/12/01]
|
| 89 |
+
\ProvidesPackage{cvpr}[2021/08/23 Example LaTex class for IEEE CVPR]
|
| 90 |
+
|
| 91 |
+
\RequirePackage{times} % Integrate Times for here
|
| 92 |
+
\RequirePackage{cite} % Automatically ordered citations
|
| 93 |
+
\RequirePackage{xspace}
|
| 94 |
+
|
| 95 |
+
\RequirePackage{silence} % Suppress unwanted warnings
|
| 96 |
+
\hbadness=10000 \vbadness=10000 \vfuzz=30pt \hfuzz=30pt
|
| 97 |
+
\WarningFilter{latexfont}{Font shape declaration}
|
| 98 |
+
\WarningFilter{latex}{Font shape}
|
| 99 |
+
\WarningFilter{hyperref}{Token not allowed in a PDF string}
|
| 100 |
+
\WarningFilter[rebuttal]{latex}{No \author given}
|
| 101 |
+
\RequirePackage{etoolbox}
|
| 102 |
+
|
| 103 |
+
% Use modern caption package to allow for sub-figures etc.
|
| 104 |
+
% Reproduces the original CVPR/ICCV style as closely as possible.
|
| 105 |
+
\RequirePackage[format=plain,labelformat=simple,labelsep=period,font=small,compatibility=false]{caption}
|
| 106 |
+
\RequirePackage[font=footnotesize,skip=3pt,subrefformat=parens]{subcaption}
|
| 107 |
+
|
| 108 |
+
|
| 109 |
+
\newtoggle{cvprfinal} % Camera-ready version
|
| 110 |
+
\newtoggle{cvprrebuttal} % Rebuttal
|
| 111 |
+
\newtoggle{cvprpagenumbers} % Force page numbers (in camera ready)
|
| 112 |
+
\toggletrue{cvprfinal}
|
| 113 |
+
\togglefalse{cvprrebuttal}
|
| 114 |
+
\togglefalse{cvprpagenumbers}
|
| 115 |
+
\DeclareOption{review}{\togglefalse{cvprfinal}\toggletrue{cvprpagenumbers}}
|
| 116 |
+
\DeclareOption{rebuttal}{\togglefalse{cvprfinal}\toggletrue{cvprrebuttal}}
|
| 117 |
+
\DeclareOption{pagenumbers}{\toggletrue{cvprpagenumbers}}
|
| 118 |
+
\DeclareOption*{\PackageWarning{cvpr}{Unkown option `\CurrentOption'}}
|
| 119 |
+
|
| 120 |
+
\ProcessOptions\relax
|
| 121 |
+
|
| 122 |
+
% Don't warn about missing author for rebuttal
|
| 123 |
+
\iftoggle{cvprrebuttal}{%
|
| 124 |
+
\ActivateWarningFilters[rebuttal]
|
| 125 |
+
}{}
|
| 126 |
+
|
| 127 |
+
% Breaking lines for URLs in the bib
|
| 128 |
+
\RequirePackage[hyphens]{url}
|
| 129 |
+
\Urlmuskip=0mu plus 1mu\relax
|
| 130 |
+
|
| 131 |
+
|
| 132 |
+
% ---------------------------------------------------------------
|
| 133 |
+
%\input{everyshi.sty}
|
| 134 |
+
\newcommand{\@EveryShipout@Hook}{}
|
| 135 |
+
\newcommand{\@EveryShipout@AtNextHook}{}
|
| 136 |
+
\newcommand*{\EveryShipout}[1]
|
| 137 |
+
{\g@addto@macro\@EveryShipout@Hook{#1}}
|
| 138 |
+
\newcommand*{\AtNextShipout}[1]
|
| 139 |
+
{\g@addto@macro\@EveryShipout@AtNextHook{#1}}
|
| 140 |
+
\newcommand{\@EveryShipout@Shipout}{%
|
| 141 |
+
\afterassignment\@EveryShipout@Test
|
| 142 |
+
\global\setbox\@cclv= %
|
| 143 |
+
}
|
| 144 |
+
\newcommand{\@EveryShipout@Test}{%
|
| 145 |
+
\ifvoid\@cclv\relax
|
| 146 |
+
\aftergroup\@EveryShipout@Output
|
| 147 |
+
\else
|
| 148 |
+
\@EveryShipout@Output
|
| 149 |
+
\fi%
|
| 150 |
+
}
|
| 151 |
+
\newcommand{\@EveryShipout@Output}{%
|
| 152 |
+
\@EveryShipout@Hook%
|
| 153 |
+
\@EveryShipout@AtNextHook%
|
| 154 |
+
\gdef\@EveryShipout@AtNextHook{}%
|
| 155 |
+
\@EveryShipout@Org@Shipout\box\@cclv%
|
| 156 |
+
}
|
| 157 |
+
\newcommand{\@EveryShipout@Org@Shipout}{}
|
| 158 |
+
\newcommand*{\@EveryShipout@Init}{%
|
| 159 |
+
\message{ABD: EveryShipout initializing macros}%
|
| 160 |
+
\let\@EveryShipout@Org@Shipout\shipout
|
| 161 |
+
\let\shipout\@EveryShipout@Shipout
|
| 162 |
+
}
|
| 163 |
+
\AtBeginDocument{\@EveryShipout@Init}
|
| 164 |
+
|
| 165 |
+
% ---------------------------------------------------------------
|
| 166 |
+
|
| 167 |
+
|
| 168 |
+
|
| 169 |
+
\newcommand\LenToUnit[1]{#1\@gobble}
|
| 170 |
+
|
| 171 |
+
\newcommand\AtPageUpperLeft[1]{%
|
| 172 |
+
\begingroup
|
| 173 |
+
\@tempdima=0pt\relax\@tempdimb=\ESO@yoffsetI\relax
|
| 174 |
+
\put(\LenToUnit{\@tempdima},\LenToUnit{\@tempdimb}){#1}%
|
| 175 |
+
\endgroup
|
| 176 |
+
}
|
| 177 |
+
\newcommand\AtPageLowerLeft[1]{\AtPageUpperLeft{%
|
| 178 |
+
\put(0,\LenToUnit{-\paperheight}){#1}}}
|
| 179 |
+
\newcommand\AtPageCenter[1]{\AtPageUpperLeft{%
|
| 180 |
+
\put(\LenToUnit{.5\paperwidth},\LenToUnit{-.5\paperheight}){#1}}%
|
| 181 |
+
}
|
| 182 |
+
\newcommand\AtTextUpperLeft[1]{%
|
| 183 |
+
\begingroup
|
| 184 |
+
\setlength\@tempdima{1in}%
|
| 185 |
+
\ifodd\c@page%
|
| 186 |
+
\advance\@tempdima\oddsidemargin%
|
| 187 |
+
\else%
|
| 188 |
+
\advance\@tempdima\evensidemargin%
|
| 189 |
+
\fi%
|
| 190 |
+
\@tempdimb=\ESO@yoffsetI\relax\advance\@tempdimb-1in\relax%
|
| 191 |
+
\advance\@tempdimb-\topmargin%
|
| 192 |
+
\advance\@tempdimb-\headheight\advance\@tempdimb-\headsep%
|
| 193 |
+
\put(\LenToUnit{\@tempdima},\LenToUnit{\@tempdimb}){#1}%
|
| 194 |
+
\endgroup
|
| 195 |
+
}
|
| 196 |
+
\newcommand\AtTextLowerLeft[1]{\AtTextUpperLeft{%
|
| 197 |
+
\put(0,\LenToUnit{-\textheight}){#1}}}
|
| 198 |
+
\newcommand\AtTextCenter[1]{\AtTextUpperLeft{%
|
| 199 |
+
\put(\LenToUnit{.5\textwidth},\LenToUnit{-.5\textheight}){#1}}}
|
| 200 |
+
\newcommand{\ESO@HookI}{} \newcommand{\ESO@HookII}{}
|
| 201 |
+
\newcommand{\ESO@HookIII}{}
|
| 202 |
+
\newcommand{\AddToShipoutPicture}{%
|
| 203 |
+
\@ifstar{\g@addto@macro\ESO@HookII}{\g@addto@macro\ESO@HookI}}
|
| 204 |
+
\newcommand{\ClearShipoutPicture}{\global\let\ESO@HookI\@empty}
|
| 205 |
+
\newcommand\ESO@isMEMOIR[1]{}
|
| 206 |
+
\@ifclassloaded{memoir}{\renewcommand\ESO@isMEMOIR[1]{#1}}{}
|
| 207 |
+
\newcommand{\@ShipoutPicture}{%
|
| 208 |
+
\bgroup
|
| 209 |
+
\@tempswafalse%
|
| 210 |
+
\ifx\ESO@HookI\@empty\else\@tempswatrue\fi%
|
| 211 |
+
\ifx\ESO@HookII\@empty\else\@tempswatrue\fi%
|
| 212 |
+
\ifx\ESO@HookIII\@empty\else\@tempswatrue\fi%
|
| 213 |
+
\if@tempswa%
|
| 214 |
+
\@tempdima=1in\@tempdimb=-\@tempdima%
|
| 215 |
+
\advance\@tempdimb\ESO@yoffsetI%
|
| 216 |
+
\ESO@isMEMOIR{%
|
| 217 |
+
\advance\@tempdima\trimedge%
|
| 218 |
+
\advance\@tempdima\paperwidth%
|
| 219 |
+
\advance\@tempdima-\stockwidth%
|
| 220 |
+
\if@twoside\ifodd\c@page\else%
|
| 221 |
+
\advance\@tempdima-2\trimedge%
|
| 222 |
+
\advance\@tempdima-\paperwidth%
|
| 223 |
+
\advance\@tempdima\stockwidth%
|
| 224 |
+
\fi\fi%
|
| 225 |
+
\advance\@tempdimb\trimtop}%
|
| 226 |
+
\unitlength=1pt%
|
| 227 |
+
\global\setbox\@cclv\vbox{%
|
| 228 |
+
\vbox{\let\protect\relax
|
| 229 |
+
\pictur@(0,0)(\strip@pt\@tempdima,\strip@pt\@tempdimb)%
|
| 230 |
+
\ESO@HookIII\ESO@HookI\ESO@HookII%
|
| 231 |
+
\global\let\ESO@HookII\@empty%
|
| 232 |
+
\endpicture}%
|
| 233 |
+
\nointerlineskip%
|
| 234 |
+
\box\@cclv}%
|
| 235 |
+
\fi
|
| 236 |
+
\egroup
|
| 237 |
+
}
|
| 238 |
+
\EveryShipout{\@ShipoutPicture}
|
| 239 |
+
\RequirePackage{keyval}
|
| 240 |
+
\newif\ifESO@dvips\ESO@dvipsfalse \newif\ifESO@grid\ESO@gridfalse
|
| 241 |
+
\newif\ifESO@texcoord\ESO@texcoordfalse
|
| 242 |
+
\newcommand*\ESO@gridunitname{}
|
| 243 |
+
\newcommand*\ESO@gridunit{}
|
| 244 |
+
\newcommand*\ESO@labelfactor{}
|
| 245 |
+
\newcommand*\ESO@griddelta{}\newcommand*\ESO@griddeltaY{}
|
| 246 |
+
\newcommand*\ESO@gridDelta{}\newcommand*\ESO@gridDeltaY{}
|
| 247 |
+
\newcommand*\ESO@gridcolor{}
|
| 248 |
+
\newcommand*\ESO@subgridcolor{}
|
| 249 |
+
\newcommand*\ESO@subgridstyle{dotted}% ???
|
| 250 |
+
\newcommand*\ESO@gap{}
|
| 251 |
+
\newcommand*\ESO@yoffsetI{}\newcommand*\ESO@yoffsetII{}
|
| 252 |
+
\newcommand*\ESO@gridlines{\thinlines}
|
| 253 |
+
\newcommand*\ESO@subgridlines{\thinlines}
|
| 254 |
+
\newcommand*\ESO@hline[1]{\ESO@subgridlines\line(1,0){#1}}
|
| 255 |
+
\newcommand*\ESO@vline[1]{\ESO@subgridlines\line(0,1){#1}}
|
| 256 |
+
\newcommand*\ESO@Hline[1]{\ESO@gridlines\line(1,0){#1}}
|
| 257 |
+
\newcommand*\ESO@Vline[1]{\ESO@gridlines\line(0,1){#1}}
|
| 258 |
+
\newcommand\ESO@fcolorbox[4][]{\fbox{#4}}
|
| 259 |
+
\newcommand\ESO@color[1]{}
|
| 260 |
+
\newcommand\ESO@colorbox[3][]{%
|
| 261 |
+
\begingroup
|
| 262 |
+
\fboxrule=0pt\fbox{#3}%
|
| 263 |
+
\endgroup
|
| 264 |
+
}
|
| 265 |
+
\newcommand\gridSetup[6][]{%
|
| 266 |
+
\edef\ESO@gridunitname{#1}\edef\ESO@gridunit{#2}
|
| 267 |
+
\edef\ESO@labelfactor{#3}\edef\ESO@griddelta{#4}
|
| 268 |
+
\edef\ESO@gridDelta{#5}\edef\ESO@gap{#6}}
|
| 269 |
+
\define@key{ESO}{texcoord}[true]{\csname ESO@texcoord#1\endcsname}
|
| 270 |
+
\define@key{ESO}{pscoord}[true]{\csname @tempswa#1\endcsname
|
| 271 |
+
\if@tempswa\ESO@texcoordfalse\else\ESO@texcoordtrue\fi}
|
| 272 |
+
\define@key{ESO}{dvips}[true]{\csname ESO@dvips#1\endcsname}
|
| 273 |
+
\define@key{ESO}{grid}[true]{\csname ESO@grid#1\endcsname
|
| 274 |
+
\setkeys{ESO}{gridcolor=black,subgridcolor=black}}
|
| 275 |
+
\define@key{ESO}{colorgrid}[true]{\csname ESO@grid#1\endcsname
|
| 276 |
+
\setkeys{ESO}{gridcolor=red,subgridcolor=green}}
|
| 277 |
+
\define@key{ESO}{gridcolor}{\def\ESO@gridcolor{#1}}
|
| 278 |
+
\define@key{ESO}{subgridcolor}{\def\ESO@subgridcolor{#1}}
|
| 279 |
+
\define@key{ESO}{subgridstyle}{\def\ESO@subgridstyle{#1}}%
|
| 280 |
+
\define@key{ESO}{gridunit}{%
|
| 281 |
+
\def\@tempa{#1}
|
| 282 |
+
\def\@tempb{bp}
|
| 283 |
+
\ifx\@tempa\@tempb
|
| 284 |
+
\gridSetup[\@tempa]{1bp}{1}{10}{50}{2}
|
| 285 |
+
\else
|
| 286 |
+
\def\@tempb{pt}
|
| 287 |
+
\ifx\@tempa\@tempb
|
| 288 |
+
\gridSetup[\@tempa]{1pt}{1}{10}{50}{2}
|
| 289 |
+
\else
|
| 290 |
+
\def\@tempb{in}
|
| 291 |
+
\ifx\@tempa\@tempb
|
| 292 |
+
\gridSetup[\@tempa]{.1in}{.1}{2}{10}{.5}
|
| 293 |
+
\else
|
| 294 |
+
\gridSetup[mm]{1mm}{1}{5}{20}{1}
|
| 295 |
+
\fi
|
| 296 |
+
\fi
|
| 297 |
+
\fi
|
| 298 |
+
}
|
| 299 |
+
%\setkeys{ESO}{subgridstyle=solid,pscoord=true,gridunit=mm}
|
| 300 |
+
% \def\ProcessOptionsWithKV#1{%
|
| 301 |
+
% \let\@tempc\@empty
|
| 302 |
+
% \@for\CurrentOption:=\@classoptionslist\do{%
|
| 303 |
+
% \@ifundefined{KV@#1@\CurrentOption}%
|
| 304 |
+
% {}{\edef\@tempc{\@tempc,\CurrentOption,}}}%
|
| 305 |
+
% \edef\@tempc{%
|
| 306 |
+
% \noexpand\setkeys{#1}{\@tempc\@ptionlist{\@currname.\@currext}}
|
| 307 |
+
% }%
|
| 308 |
+
% \@tempc
|
| 309 |
+
% \AtEndOfPackage{\let\@unprocessedoptions\relax}}%
|
| 310 |
+
%\ProcessOptionsWithKV{ESO}%
|
| 311 |
+
\newcommand\ESO@div[2]{%
|
| 312 |
+
\@tempdima=#1\relax\@tempdimb=\ESO@gridunit\relax
|
| 313 |
+
\@tempdimb=#2\@tempdimb\divide\@tempdima by \@tempdimb%
|
| 314 |
+
\@tempcnta\@tempdima\advance\@tempcnta\@ne}
|
| 315 |
+
\AtBeginDocument{%
|
| 316 |
+
\IfFileExists{color.sty}
|
| 317 |
+
{%
|
| 318 |
+
\RequirePackage{color}
|
| 319 |
+
\let\ESO@color=\color\let\ESO@colorbox=\colorbox
|
| 320 |
+
\let\ESO@fcolorbox=\fcolorbox
|
| 321 |
+
}{}
|
| 322 |
+
\@ifundefined{Gin@driver}{}%
|
| 323 |
+
{%
|
| 324 |
+
\ifx\Gin@driver\@empty\else%
|
| 325 |
+
\filename@parse{\Gin@driver}\def\reserved@a{dvips}%
|
| 326 |
+
\ifx\filename@base\reserved@a\ESO@dvipstrue\fi%
|
| 327 |
+
\fi
|
| 328 |
+
}%
|
| 329 |
+
\ifx\pdfoutput\undefined\else
|
| 330 |
+
\ifx\pdfoutput\relax\else
|
| 331 |
+
\ifcase\pdfoutput\else
|
| 332 |
+
\ESO@dvipsfalse%
|
| 333 |
+
\fi
|
| 334 |
+
\fi
|
| 335 |
+
\fi
|
| 336 |
+
\ifESO@dvips\def\@tempb{eepic}\else\def\@tempb{epic}\fi
|
| 337 |
+
\def\@tempa{dotted}%\def\ESO@gap{\LenToUnit{6\@wholewidth}}%
|
| 338 |
+
\ifx\@tempa\ESO@subgridstyle
|
| 339 |
+
\IfFileExists{\@tempb.sty}%
|
| 340 |
+
{%
|
| 341 |
+
\RequirePackage{\@tempb}
|
| 342 |
+
\renewcommand*\ESO@hline[1]{\ESO@subgridlines\dottedline{\ESO@gap}%
|
| 343 |
+
(0,0)(##1,0)}
|
| 344 |
+
\renewcommand*\ESO@vline[1]{\ESO@subgridlines\dottedline{\ESO@gap}%
|
| 345 |
+
(0,0)(0,##1)}
|
| 346 |
+
}{}
|
| 347 |
+
\else
|
| 348 |
+
\ifx\ESO@gridcolor\ESO@subgridcolor%
|
| 349 |
+
\renewcommand*\ESO@gridlines{\thicklines}
|
| 350 |
+
\fi
|
| 351 |
+
\fi
|
| 352 |
+
}
|
| 353 |
+
\ifESO@texcoord
|
| 354 |
+
\def\ESO@yoffsetI{0pt}\def\ESO@yoffsetII{-\paperheight}
|
| 355 |
+
\edef\ESO@griddeltaY{-\ESO@griddelta}\edef\ESO@gridDeltaY{-\ESO@gridDelta}
|
| 356 |
+
\else
|
| 357 |
+
\def\ESO@yoffsetI{\paperheight}\def\ESO@yoffsetII{0pt}
|
| 358 |
+
\edef\ESO@griddeltaY{\ESO@griddelta}\edef\ESO@gridDeltaY{\ESO@gridDelta}
|
| 359 |
+
\fi
|
| 360 |
+
\newcommand\ESO@gridpicture{%
|
| 361 |
+
\begingroup
|
| 362 |
+
\setlength\unitlength{\ESO@gridunit}%
|
| 363 |
+
\ESO@color{\ESO@subgridcolor}%
|
| 364 |
+
\ESO@div{\paperheight}{\ESO@griddelta}%
|
| 365 |
+
\multiput(0,0)(0,\ESO@griddeltaY){\@tempcnta}%
|
| 366 |
+
{\ESO@hline{\LenToUnit{\paperwidth}}}%
|
| 367 |
+
\ESO@div{\paperwidth}{\ESO@griddelta}%
|
| 368 |
+
\multiput(0,\LenToUnit{\ESO@yoffsetII})(\ESO@griddelta,0){\@tempcnta}%
|
| 369 |
+
{\ESO@vline{\LenToUnit{\paperheight}}}%
|
| 370 |
+
\ESO@color{\ESO@gridcolor}%
|
| 371 |
+
\ESO@div{\paperheight}{\ESO@gridDelta}%
|
| 372 |
+
\multiput(0,0)(0,\ESO@gridDeltaY){\@tempcnta}%
|
| 373 |
+
{\ESO@Hline{\LenToUnit{\paperwidth}}}%
|
| 374 |
+
\ESO@div{\paperwidth}{\ESO@gridDelta}%
|
| 375 |
+
\multiput(0,\LenToUnit{\ESO@yoffsetII})(\ESO@gridDelta,0){\@tempcnta}%
|
| 376 |
+
{\ESO@Vline{\LenToUnit{\paperheight}}}%
|
| 377 |
+
\fontsize{10}{12}\normalfont%
|
| 378 |
+
\ESO@div{\paperwidth}{\ESO@gridDelta}%
|
| 379 |
+
\multiput(0,\ESO@gridDeltaY)(\ESO@gridDelta,0){\@tempcnta}{%
|
| 380 |
+
\@tempcntb=\@tempcnta\advance\@tempcntb-\@multicnt%
|
| 381 |
+
\ifnum\@tempcntb>1\relax
|
| 382 |
+
\multiply\@tempcntb by \ESO@gridDelta\relax%
|
| 383 |
+
\@tempdima=\@tempcntb sp\@tempdima=\ESO@labelfactor\@tempdima%
|
| 384 |
+
\@tempcntb=\@tempdima%
|
| 385 |
+
\makebox(0,0)[c]{\ESO@colorbox{white}{\the\@tempcntb}}%
|
| 386 |
+
\fi}%
|
| 387 |
+
\ifx\ESO@gridunitname\@empty\def\@tempa{0}\else\def\@tempa{1}\fi%
|
| 388 |
+
\ESO@div{\paperheight}{\ESO@gridDelta}%
|
| 389 |
+
\multiput(\ESO@gridDelta,0)(0,\ESO@gridDeltaY){\@tempcnta}{%
|
| 390 |
+
\@tempcntb=\@tempcnta\advance\@tempcntb-\@multicnt%
|
| 391 |
+
\ifnum\@tempcntb>\@tempa\relax
|
| 392 |
+
\multiply\@tempcntb by \ESO@gridDelta\relax%
|
| 393 |
+
\@tempdima=\@tempcntb sp\@tempdima=\ESO@labelfactor\@tempdima%
|
| 394 |
+
\@tempcntb=\@tempdima%
|
| 395 |
+
\makebox(0,0)[c]{\ESO@colorbox{white}{\the\@tempcntb}}%
|
| 396 |
+
\fi
|
| 397 |
+
}%
|
| 398 |
+
\ifx\ESO@gridunitname\@empty\else%
|
| 399 |
+
\thicklines\fboxrule=\@wholewidth%
|
| 400 |
+
\put(\ESO@gridDelta,\ESO@gridDeltaY){\makebox(0,0)[c]{%
|
| 401 |
+
\ESO@fcolorbox{\ESO@gridcolor}{white}{%
|
| 402 |
+
\textbf{\ESO@gridunitname}}}}%
|
| 403 |
+
\fi
|
| 404 |
+
\normalcolor%
|
| 405 |
+
\endgroup
|
| 406 |
+
}
|
| 407 |
+
\ifESO@grid\g@addto@macro\ESO@HookIII{\ESO@gridpicture}\fi
|
| 408 |
+
% ---------------------------------------------------------------
|
| 409 |
+
|
| 410 |
+
|
| 411 |
+
|
| 412 |
+
|
| 413 |
+
\typeout{CVPR 8.5 x 11-Inch Proceedings Style `cvpr.sty'.}
|
| 414 |
+
|
| 415 |
+
% ten point helvetica bold required for captions
|
| 416 |
+
% eleven point times bold required for second-order headings
|
| 417 |
+
% in some sites the name of the fonts may differ,
|
| 418 |
+
% change the name here:
|
| 419 |
+
\font\cvprtenhv = phvb at 8pt % *** IF THIS FAILS, SEE cvpr.sty ***
|
| 420 |
+
\font\elvbf = ptmb scaled 1100
|
| 421 |
+
|
| 422 |
+
% If the above lines give an error message, try to comment them and
|
| 423 |
+
% uncomment these:
|
| 424 |
+
%\font\cvprtenhv = phvb7t at 8pt
|
| 425 |
+
%\font\elvbf = ptmb7t scaled 1100
|
| 426 |
+
|
| 427 |
+
% set dimensions of columns, gap between columns, and paragraph indent
|
| 428 |
+
\setlength{\textheight}{8.875in}
|
| 429 |
+
\setlength{\textwidth}{6.875in}
|
| 430 |
+
\setlength{\columnsep}{0.3125in}
|
| 431 |
+
\setlength{\topmargin}{0in}
|
| 432 |
+
\setlength{\headheight}{0in}
|
| 433 |
+
\setlength{\headsep}{0in}
|
| 434 |
+
\setlength{\parindent}{1pc}
|
| 435 |
+
\setlength{\oddsidemargin}{-.304in}
|
| 436 |
+
\setlength{\evensidemargin}{-.304in}
|
| 437 |
+
|
| 438 |
+
|
| 439 |
+
% memento from size10.clo
|
| 440 |
+
% \normalsize{\@setfontsize\normalsize\@xpt\@xiipt}
|
| 441 |
+
% \small{\@setfontsize\small\@ixpt{11}}
|
| 442 |
+
% \footnotesize{\@setfontsize\footnotesize\@viiipt{9.5}}
|
| 443 |
+
% \scriptsize{\@setfontsize\scriptsize\@viipt\@viiipt}
|
| 444 |
+
% \tiny{\@setfontsize\tiny\@vpt\@vipt}
|
| 445 |
+
% \large{\@setfontsize\large\@xiipt{14}}
|
| 446 |
+
% \Large{\@setfontsize\Large\@xivpt{18}}
|
| 447 |
+
% \LARGE{\@setfontsize\LARGE\@xviipt{22}}
|
| 448 |
+
% \huge{\@setfontsize\huge\@xxpt{25}}
|
| 449 |
+
% \Huge{\@setfontsize\Huge\@xxvpt{30}}
|
| 450 |
+
|
| 451 |
+
|
| 452 |
+
% Suppress page numbers when the appropriate option is given
|
| 453 |
+
\iftoggle{cvprpagenumbers}{}{%
|
| 454 |
+
\pagestyle{empty}
|
| 455 |
+
}
|
| 456 |
+
|
| 457 |
+
\AtBeginDocument{%
|
| 458 |
+
% Print an error if document class other than article is used
|
| 459 |
+
\@ifclassloaded{article}{}{%
|
| 460 |
+
\PackageError{cvpr}{Package only meant to be used with document class `article'}{Change document class to `article'.}
|
| 461 |
+
}
|
| 462 |
+
% Print a warning if incorrect options for article are specified
|
| 463 |
+
\@ifclasswith{article}{10pt}{}{%
|
| 464 |
+
\PackageWarningNoLine{cvpr}{Incorrect font size specified - CVPR requires 10-point fonts. Please load document class `article' with `10pt' option}
|
| 465 |
+
}
|
| 466 |
+
\@ifclasswith{article}{twocolumn}{}{%
|
| 467 |
+
\PackageWarningNoLine{cvpr}{Single column document - CVPR requires papers to have two-column layout. Please load document class `article' with `twocolumn' option}
|
| 468 |
+
}
|
| 469 |
+
\@ifclasswith{article}{letterpaper}{}{%
|
| 470 |
+
\PackageWarningNoLine{cvpr}{Incorrect paper size - CVPR uses paper size `letter'. Please load document class `article' with `letterpaper' option}
|
| 471 |
+
}
|
| 472 |
+
% Print a warning if hyperref is not loaded and/or if the pagebackref option is missing
|
| 473 |
+
\iftoggle{cvprfinal}{%
|
| 474 |
+
\@ifpackageloaded{hyperref}{}{%
|
| 475 |
+
\PackageWarningNoLine{cvpr}{Package `hyperref' is not loaded, but highly recommended for camera-ready version}
|
| 476 |
+
}
|
| 477 |
+
}{%
|
| 478 |
+
\@ifpackageloaded{hyperref}{
|
| 479 |
+
\@ifpackagewith{hyperref}{pagebackref}{}{
|
| 480 |
+
\PackageWarningNoLine{cvpr}{Package `hyperref' is not loaded with option `pagebackref', which is strongly recommended for review version}
|
| 481 |
+
}
|
| 482 |
+
}{%
|
| 483 |
+
\PackageWarningNoLine{cvpr}{Package `hyperref' is not loaded, but strongly recommended for review version}
|
| 484 |
+
}
|
| 485 |
+
}
|
| 486 |
+
}
|
| 487 |
+
|
| 488 |
+
\def\@maketitle
|
| 489 |
+
{
|
| 490 |
+
\newpage
|
| 491 |
+
\null
|
| 492 |
+
\iftoggle{cvprrebuttal}{\vspace*{-.3in}}{\vskip .375in}
|
| 493 |
+
\begin{center}
|
| 494 |
+
% smaller title font only for rebuttal
|
| 495 |
+
\iftoggle{cvprrebuttal}{{\large \bf \@title \par}}{{\Large \bf \@title \par}}
|
| 496 |
+
% additional two empty lines at the end of the title
|
| 497 |
+
\iftoggle{cvprrebuttal}{\vspace*{-22pt}}{\vspace*{24pt}}
|
| 498 |
+
{
|
| 499 |
+
\large
|
| 500 |
+
\lineskip .5em
|
| 501 |
+
\begin{tabular}[t]{c}
|
| 502 |
+
\iftoggle{cvprfinal}{
|
| 503 |
+
\@author
|
| 504 |
+
}{
|
| 505 |
+
\iftoggle{cvprrebuttal}{}{
|
| 506 |
+
Anonymous \confName~submission\\
|
| 507 |
+
\vspace*{1pt}\\
|
| 508 |
+
Paper ID \cvprPaperID
|
| 509 |
+
}
|
| 510 |
+
}
|
| 511 |
+
\end{tabular}
|
| 512 |
+
\par
|
| 513 |
+
}
|
| 514 |
+
% additional small space at the end of the author name
|
| 515 |
+
\vskip .5em
|
| 516 |
+
% additional empty line at the end of the title block
|
| 517 |
+
\vspace*{12pt}
|
| 518 |
+
\end{center}
|
| 519 |
+
}
|
| 520 |
+
|
| 521 |
+
\def\abstract
|
| 522 |
+
{%
|
| 523 |
+
% Suppress page numbers when the appropriate option is given
|
| 524 |
+
\iftoggle{cvprpagenumbers}{}{%
|
| 525 |
+
\thispagestyle{empty}
|
| 526 |
+
}
|
| 527 |
+
\centerline{\large\bf Abstract}%
|
| 528 |
+
\vspace*{12pt}%
|
| 529 |
+
\it%
|
| 530 |
+
}
|
| 531 |
+
|
| 532 |
+
\def\endabstract
|
| 533 |
+
{
|
| 534 |
+
% additional empty line at the end of the abstract
|
| 535 |
+
\vspace*{12pt}
|
| 536 |
+
}
|
| 537 |
+
|
| 538 |
+
\def\affiliation#1{\gdef\@affiliation{#1}} \gdef\@affiliation{}
|
| 539 |
+
|
| 540 |
+
% correct heading spacing and type
|
| 541 |
+
\def\cvprsection{\@startsection {section}{1}{\z@}
|
| 542 |
+
{10pt plus 2pt minus 2pt}{7pt} {\large\bf}}
|
| 543 |
+
\def\cvprssect#1{\cvprsection*{#1}}
|
| 544 |
+
\def\cvprsect#1{\cvprsection{\hskip -1em.~#1}}
|
| 545 |
+
\def\section{\@ifstar\cvprssect\cvprsect}
|
| 546 |
+
|
| 547 |
+
\def\cvprsubsection{\@startsection {subsection}{2}{\z@}
|
| 548 |
+
{8pt plus 2pt minus 2pt}{6pt} {\elvbf}}
|
| 549 |
+
\def\cvprssubsect#1{\cvprsubsection*{#1}}
|
| 550 |
+
\def\cvprsubsect#1{\cvprsubsection{\hskip -1em.~#1}}
|
| 551 |
+
\def\subsection{\@ifstar\cvprssubsect\cvprsubsect}
|
| 552 |
+
|
| 553 |
+
%% --------- Page background marks: Ruler and confidentiality
|
| 554 |
+
|
| 555 |
+
% ----- define vruler
|
| 556 |
+
\makeatletter
|
| 557 |
+
\newbox\cvprrulerbox
|
| 558 |
+
\newcount\cvprrulercount
|
| 559 |
+
\newdimen\cvprruleroffset
|
| 560 |
+
\newdimen\cv@lineheight
|
| 561 |
+
\newdimen\cv@boxheight
|
| 562 |
+
\newbox\cv@tmpbox
|
| 563 |
+
\newcount\cv@refno
|
| 564 |
+
\newcount\cv@tot
|
| 565 |
+
% NUMBER with left flushed zeros \fillzeros[<WIDTH>]<NUMBER>
|
| 566 |
+
\newcount\cv@tmpc@ \newcount\cv@tmpc
|
| 567 |
+
\def\fillzeros[#1]#2{\cv@tmpc@=#2\relax\ifnum\cv@tmpc@<0\cv@tmpc@=-\cv@tmpc@\fi
|
| 568 |
+
\cv@tmpc=1 %
|
| 569 |
+
\loop\ifnum\cv@tmpc@<10 \else \divide\cv@tmpc@ by 10 \advance\cv@tmpc by 1 \fi
|
| 570 |
+
\ifnum\cv@tmpc@=10\relax\cv@tmpc@=11\relax\fi \ifnum\cv@tmpc@>10 \repeat
|
| 571 |
+
\ifnum#2<0\advance\cv@tmpc1\relax-\fi
|
| 572 |
+
\loop\ifnum\cv@tmpc<#1\relax0\advance\cv@tmpc1\relax\fi \ifnum\cv@tmpc<#1 \repeat
|
| 573 |
+
\cv@tmpc@=#2\relax\ifnum\cv@tmpc@<0\cv@tmpc@=-\cv@tmpc@\fi \relax\the\cv@tmpc@}%
|
| 574 |
+
% \makevruler[<SCALE>][<INITIAL_COUNT>][<STEP>][<DIGITS>][<HEIGHT>]
|
| 575 |
+
\def\makevruler[#1][#2][#3][#4][#5]{\begingroup\offinterlineskip
|
| 576 |
+
\textheight=#5\vbadness=10000\vfuzz=120ex\overfullrule=0pt%
|
| 577 |
+
\global\setbox\cvprrulerbox=\vbox to \textheight{%
|
| 578 |
+
{\parskip=0pt\hfuzz=150em\cv@boxheight=\textheight
|
| 579 |
+
\cv@lineheight=#1\global\cvprrulercount=#2%
|
| 580 |
+
\cv@tot\cv@boxheight\divide\cv@tot\cv@lineheight\advance\cv@tot2%
|
| 581 |
+
\cv@refno1\vskip-\cv@lineheight\vskip1ex%
|
| 582 |
+
\loop\setbox\cv@tmpbox=\hbox to0cm{{\cvprtenhv\hfil\fillzeros[#4]\cvprrulercount}}%
|
| 583 |
+
\ht\cv@tmpbox\cv@lineheight\dp\cv@tmpbox0pt\box\cv@tmpbox\break
|
| 584 |
+
\advance\cv@refno1\global\advance\cvprrulercount#3\relax
|
| 585 |
+
\ifnum\cv@refno<\cv@tot\repeat}}\endgroup}%
|
| 586 |
+
\makeatother
|
| 587 |
+
% ----- end of vruler
|
| 588 |
+
|
| 589 |
+
% \makevruler[<SCALE>][<INITIAL_COUNT>][<STEP>][<DIGITS>][<HEIGHT>]
|
| 590 |
+
\def\cvprruler#1{\makevruler[12pt][#1][1][3][0.993\textheight]\usebox{\cvprrulerbox}}
|
| 591 |
+
\AddToShipoutPicture{%
|
| 592 |
+
\iftoggle{cvprfinal}{
|
| 593 |
+
}{
|
| 594 |
+
\cvprruleroffset=\textheight
|
| 595 |
+
\advance\cvprruleroffset by -3.7pt
|
| 596 |
+
\color[rgb]{.5,.5,1}
|
| 597 |
+
\AtTextUpperLeft{%
|
| 598 |
+
\put(\LenToUnit{-35pt},\LenToUnit{-\cvprruleroffset}){%left ruler
|
| 599 |
+
\cvprruler{\cvprrulercount}}
|
| 600 |
+
%\put(\LenToUnit{\textwidth\kern 30pt},\LenToUnit{-\cvprruleroffset})
|
| 601 |
+
\put(\LenToUnit{\dimexpr \textwidth+30pt},\LenToUnit{-\cvprruleroffset}){%right ruler
|
| 602 |
+
\cvprruler{\cvprrulercount}}
|
| 603 |
+
}
|
| 604 |
+
\def\pid{\parbox{1in}{\begin{center}\bf\sf{\small \confName}\\\#\cvprPaperID\end{center}}}
|
| 605 |
+
\AtTextUpperLeft{%paperID in corners
|
| 606 |
+
\put(\LenToUnit{-65pt},\LenToUnit{45pt}){\pid}
|
| 607 |
+
\put(\LenToUnit{\textwidth\kern-8pt},\LenToUnit{45pt}){\pid}
|
| 608 |
+
}
|
| 609 |
+
\AtTextUpperLeft{%confidential
|
| 610 |
+
\put(0,\LenToUnit{1cm}){\parbox{\textwidth}{\centering\cvprtenhv
|
| 611 |
+
\confName~\confYear~Submission \#\cvprPaperID. CONFIDENTIAL REVIEW COPY. DO NOT DISTRIBUTE.}}
|
| 612 |
+
}
|
| 613 |
+
}
|
| 614 |
+
}
|
| 615 |
+
|
| 616 |
+
%%% Make figure placement a little more predictable.
|
| 617 |
+
% We trust the user to move figures if this results
|
| 618 |
+
% in ugliness.
|
| 619 |
+
% Minimize bad page breaks at figures
|
| 620 |
+
\renewcommand{\textfraction}{0.01}
|
| 621 |
+
\renewcommand{\floatpagefraction}{0.99}
|
| 622 |
+
\renewcommand{\topfraction}{0.99}
|
| 623 |
+
\renewcommand{\bottomfraction}{0.99}
|
| 624 |
+
\renewcommand{\dblfloatpagefraction}{0.99}
|
| 625 |
+
\renewcommand{\dbltopfraction}{0.99}
|
| 626 |
+
\setcounter{totalnumber}{99}
|
| 627 |
+
\setcounter{topnumber}{99}
|
| 628 |
+
\setcounter{bottomnumber}{99}
|
| 629 |
+
|
| 630 |
+
% Add a period to the end of an abbreviation unless there's one
|
| 631 |
+
% already, then \xspace.
|
| 632 |
+
\makeatletter
|
| 633 |
+
\DeclareRobustCommand\onedot{\futurelet\@let@token\@onedot}
|
| 634 |
+
\def\@onedot{\ifx\@let@token.\else.\null\fi\xspace}
|
| 635 |
+
|
| 636 |
+
\def\eg{\emph{e.g}\onedot} \def\Eg{\emph{E.g}\onedot}
|
| 637 |
+
\def\ie{\emph{i.e}\onedot} \def\Ie{\emph{I.e}\onedot}
|
| 638 |
+
\def\cf{\emph{cf}\onedot} \def\Cf{\emph{Cf}\onedot}
|
| 639 |
+
\def\etc{\emph{etc}\onedot} \def\vs{\emph{vs}\onedot}
|
| 640 |
+
\def\wrt{w.r.t\onedot} \def\dof{d.o.f\onedot}
|
| 641 |
+
\def\iid{i.i.d\onedot} \def\wolog{w.l.o.g\onedot}
|
| 642 |
+
\def\etal{\emph{et al}\onedot}
|
| 643 |
+
\makeatother
|
| 644 |
+
|
| 645 |
+
% ---------------------------------------------------------------
|
paper-tex/egbib.bib
ADDED
|
@@ -0,0 +1,592 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
@String(PAMI = {IEEE Trans. Pattern Anal. Mach. Intell.})
|
| 2 |
+
@String(IJCV = {Int. J. Comput. Vis.})
|
| 3 |
+
@String(CVPR= {IEEE Conf. Comput. Vis. Pattern Recog.})
|
| 4 |
+
@String(ICCV= {Int. Conf. Comput. Vis.})
|
| 5 |
+
@String(ECCV= {Eur. Conf. Comput. Vis.})
|
| 6 |
+
@String(NIPS= {Adv. Neural Inform. Process. Syst.})
|
| 7 |
+
@String(ICPR = {Int. Conf. Pattern Recog.})
|
| 8 |
+
@String(BMVC= {Brit. Mach. Vis. Conf.})
|
| 9 |
+
@String(TOG= {ACM Trans. Graph.})
|
| 10 |
+
@String(TIP = {IEEE Trans. Image Process.})
|
| 11 |
+
@String(TVCG = {IEEE Trans. Vis. Comput. Graph.})
|
| 12 |
+
@String(TMM = {IEEE Trans. Multimedia})
|
| 13 |
+
@String(ACMMM= {ACM Int. Conf. Multimedia})
|
| 14 |
+
@String(ICME = {Int. Conf. Multimedia and Expo})
|
| 15 |
+
@String(ICASSP= {ICASSP})
|
| 16 |
+
@String(ICIP = {IEEE Int. Conf. Image Process.})
|
| 17 |
+
@String(ACCV = {ACCV})
|
| 18 |
+
@String(ICLR = {Int. Conf. Learn. Represent.})
|
| 19 |
+
@String(IJCAI = {IJCAI})
|
| 20 |
+
@String(PR = {Pattern Recognition})
|
| 21 |
+
@String(AAAI = {AAAI})
|
| 22 |
+
@String(CVPRW= {IEEE Conf. Comput. Vis. Pattern Recog. Worksh.})
|
| 23 |
+
@String(CSVT = {IEEE Trans. Circuit Syst. Video Technol.})
|
| 24 |
+
|
| 25 |
+
@String(SPL = {IEEE Sign. Process. Letters})
|
| 26 |
+
@String(VR = {Vis. Res.})
|
| 27 |
+
@String(JOV = {J. Vis.})
|
| 28 |
+
@String(TVC = {The Vis. Comput.})
|
| 29 |
+
@String(JCST = {J. Comput. Sci. Tech.})
|
| 30 |
+
@String(CGF = {Comput. Graph. Forum})
|
| 31 |
+
@String(CVM = {Computational Visual Media})
|
| 32 |
+
|
| 33 |
+
|
| 34 |
+
@String(PAMI = {IEEE TPAMI})
|
| 35 |
+
@String(IJCV = {IJCV})
|
| 36 |
+
@String(CVPR = {CVPR})
|
| 37 |
+
@String(ICCV = {ICCV})
|
| 38 |
+
@String(ECCV = {ECCV})
|
| 39 |
+
@String(NIPS = {NeurIPS})
|
| 40 |
+
@String(ICPR = {ICPR})
|
| 41 |
+
@String(BMVC = {BMVC})
|
| 42 |
+
@String(TOG = {ACM TOG})
|
| 43 |
+
@String(TIP = {IEEE TIP})
|
| 44 |
+
@String(TVCG = {IEEE TVCG})
|
| 45 |
+
@String(TCSVT = {IEEE TCSVT})
|
| 46 |
+
@String(TMM = {IEEE TMM})
|
| 47 |
+
@String(ACMMM = {ACM MM})
|
| 48 |
+
@String(ICME = {ICME})
|
| 49 |
+
@String(ICASSP= {ICASSP})
|
| 50 |
+
@String(ICIP = {ICIP})
|
| 51 |
+
@String(ACCV = {ACCV})
|
| 52 |
+
@String(ICLR = {ICLR})
|
| 53 |
+
@String(IJCAI = {IJCAI})
|
| 54 |
+
@String(PR = {PR})
|
| 55 |
+
@String(AAAI = {AAAI})
|
| 56 |
+
@String(CVPRW= {CVPRW})
|
| 57 |
+
@String(CSVT = {IEEE TCSVT})
|
| 58 |
+
|
| 59 |
+
@inproceedings{resnet,
|
| 60 |
+
title={Deep residual learning for image recognition},
|
| 61 |
+
author={He, Kaiming and Zhang, Xiangyu and Ren, Shaoqing and Sun, Jian},
|
| 62 |
+
booktitle={Proceedings of the IEEE conference on computer vision and pattern recognition},
|
| 63 |
+
pages={770--778},
|
| 64 |
+
year={2016}
|
| 65 |
+
}
|
| 66 |
+
|
| 67 |
+
@inproceedings{deeplabv3p,
|
| 68 |
+
title={Encoder-decoder with atrous separable convolution for semantic image segmentation},
|
| 69 |
+
author={Chen, Liang-Chieh and Zhu, Yukun and Papandreou, George and Schroff, Florian and Adam, Hartwig},
|
| 70 |
+
booktitle={Proceedings of the European conference on computer vision (ECCV)},
|
| 71 |
+
pages={801--818},
|
| 72 |
+
year={2018}
|
| 73 |
+
}
|
| 74 |
+
|
| 75 |
+
@article{deeplabv3,
|
| 76 |
+
title={Rethinking atrous convolution for semantic image segmentation},
|
| 77 |
+
author={Chen, Liang-Chieh and Papandreou, George and Schroff, Florian and Adam, Hartwig},
|
| 78 |
+
journal={arXiv preprint arXiv:1706.05587},
|
| 79 |
+
year={2017}
|
| 80 |
+
}
|
| 81 |
+
|
| 82 |
+
@article{deeplab,
|
| 83 |
+
title={Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs},
|
| 84 |
+
author={Chen, Liang-Chieh and Papandreou, George and Kokkinos, Iasonas and Murphy, Kevin and Yuille, Alan L},
|
| 85 |
+
journal={IEEE transactions on pattern analysis and machine intelligence},
|
| 86 |
+
volume={40},
|
| 87 |
+
number={4},
|
| 88 |
+
pages={834--848},
|
| 89 |
+
year={2017},
|
| 90 |
+
publisher={IEEE}
|
| 91 |
+
}
|
| 92 |
+
|
| 93 |
+
@inproceedings{pspnet,
|
| 94 |
+
title={Pyramid scene parsing network},
|
| 95 |
+
author={Zhao, Hengshuang and Shi, Jianping and Qi, Xiaojuan and Wang, Xiaogang and Jia, Jiaya},
|
| 96 |
+
booktitle={Proceedings of the IEEE conference on computer vision and pattern recognition},
|
| 97 |
+
pages={2881--2890},
|
| 98 |
+
year={2017}
|
| 99 |
+
}
|
| 100 |
+
|
| 101 |
+
@inproceedings{ccnet,
|
| 102 |
+
title={Ccnet: Criss-cross attention for semantic segmentation},
|
| 103 |
+
author={Huang, Zilong and Wang, Xinggang and Huang, Lichao and Huang, Chang and Wei, Yunchao and Liu, Wenyu},
|
| 104 |
+
booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
|
| 105 |
+
pages={603--612},
|
| 106 |
+
year={2019}
|
| 107 |
+
}
|
| 108 |
+
|
| 109 |
+
@inproceedings{attention,
|
| 110 |
+
title={Attention is all you need},
|
| 111 |
+
author={Vaswani, Ashish and Shazeer, Noam and Parmar, Niki and Uszkoreit, Jakob and Jones, Llion and Gomez, Aidan N and Kaiser, {\L}ukasz and Polosukhin, Illia},
|
| 112 |
+
booktitle={Advances in neural information processing systems},
|
| 113 |
+
pages={5998--6008},
|
| 114 |
+
year={2017}
|
| 115 |
+
}
|
| 116 |
+
|
| 117 |
+
@article{vit,
|
| 118 |
+
title={An image is worth 16x16 words: Transformers for image recognition at scale},
|
| 119 |
+
author={Dosovitskiy, Alexey and Beyer, Lucas and Kolesnikov, Alexander and Weissenborn, Dirk and Zhai, Xiaohua and Unterthiner, Thomas and Dehghani, Mostafa and Minderer, Matthias and Heigold, Georg and Gelly, Sylvain and others},
|
| 120 |
+
journal={arXiv preprint arXiv:2010.11929},
|
| 121 |
+
year={2020}
|
| 122 |
+
}
|
| 123 |
+
|
| 124 |
+
@article{swin,
|
| 125 |
+
title={Swin transformer: Hierarchical vision transformer using shifted windows},
|
| 126 |
+
author={Liu, Ze and Lin, Yutong and Cao, Yue and Hu, Han and Wei, Yixuan and Zhang, Zheng and Lin, Stephen and Guo, Baining},
|
| 127 |
+
journal={arXiv preprint arXiv:2103.14030},
|
| 128 |
+
year={2021}
|
| 129 |
+
}
|
| 130 |
+
|
| 131 |
+
@inproceedings{dpt,
|
| 132 |
+
title={Vision transformers for dense prediction},
|
| 133 |
+
author={Ranftl, Ren{\'e} and Bochkovskiy, Alexey and Koltun, Vladlen},
|
| 134 |
+
booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
|
| 135 |
+
pages={12179--12188},
|
| 136 |
+
year={2021}
|
| 137 |
+
}
|
| 138 |
+
|
| 139 |
+
@inproceedings{setr,
|
| 140 |
+
title={Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers},
|
| 141 |
+
author={Zheng, Sixiao and Lu, Jiachen and Zhao, Hengshuang and Zhu, Xiatian and Luo, Zekun and Wang, Yabiao and Fu, Yanwei and Feng, Jianfeng and Xiang, Tao and Torr, Philip HS and others},
|
| 142 |
+
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
|
| 143 |
+
pages={6881--6890},
|
| 144 |
+
year={2021}
|
| 145 |
+
}
|
| 146 |
+
|
| 147 |
+
@article{segformer,
|
| 148 |
+
title={SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers},
|
| 149 |
+
author={Xie, Enze and Wang, Wenhai and Yu, Zhiding and Anandkumar, Anima and Alvarez, Jose M and Luo, Ping},
|
| 150 |
+
journal={arXiv preprint arXiv:2105.15203},
|
| 151 |
+
year={2021}
|
| 152 |
+
}
|
| 153 |
+
|
| 154 |
+
@inproceedings{glnet,
|
| 155 |
+
title={Collaborative global-local networks for memory-efficient segmentation of ultra-high resolution images},
|
| 156 |
+
author={Chen, Wuyang and Jiang, Ziyu and Wang, Zhangyang and Cui, Kexin and Qian, Xiaoning},
|
| 157 |
+
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
|
| 158 |
+
pages={8924--8933},
|
| 159 |
+
year={2019}
|
| 160 |
+
}
|
| 161 |
+
|
| 162 |
+
@inproceedings{cascadepsp,
|
| 163 |
+
title={CascadePSP: toward class-agnostic and very high-resolution segmentation via global and local refinement},
|
| 164 |
+
author={Cheng, Ho Kei and Chung, Jihoon and Tai, Yu-Wing and Tang, Chi-Keung},
|
| 165 |
+
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
|
| 166 |
+
pages={8890--8899},
|
| 167 |
+
year={2020}
|
| 168 |
+
}
|
| 169 |
+
|
| 170 |
+
@inproceedings{magnet,
|
| 171 |
+
title={Progressive Semantic Segmentation},
|
| 172 |
+
author={Huynh, Chuong and Tran, Anh Tuan and Luu, Khoa and Hoai, Minh},
|
| 173 |
+
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
|
| 174 |
+
pages={16755--16764},
|
| 175 |
+
year={2021}
|
| 176 |
+
}
|
| 177 |
+
|
| 178 |
+
@article{cable_inst,
|
| 179 |
+
title={Transmission line detection in aerial images: An instance segmentation approach based on multitask neural networks},
|
| 180 |
+
author={Li, Bo and Chen, Cheng and Dong, Shiwen and Qiao, Junfeng},
|
| 181 |
+
journal={Signal Processing: Image Communication},
|
| 182 |
+
volume={96},
|
| 183 |
+
pages={116278},
|
| 184 |
+
year={2021},
|
| 185 |
+
publisher={Elsevier}
|
| 186 |
+
}
|
| 187 |
+
|
| 188 |
+
@article{pldu,
|
| 189 |
+
title={Detecting power lines in UAV images with convolutional features and structured constraints},
|
| 190 |
+
author={Zhang, Heng and Yang, Wen and Yu, Huai and Zhang, Haijian and Xia, Gui-Song},
|
| 191 |
+
journal={Remote Sensing},
|
| 192 |
+
volume={11},
|
| 193 |
+
number={11},
|
| 194 |
+
pages={1342},
|
| 195 |
+
year={2019},
|
| 196 |
+
publisher={Multidisciplinary Digital Publishing Institute}
|
| 197 |
+
}
|
| 198 |
+
|
| 199 |
+
@inproceedings{ttpla,
|
| 200 |
+
title={TTPLA: An Aerial-Image Dataset for Detection and Segmentation of Transmission Towers and Power Lines},
|
| 201 |
+
author={Abdelfattah, Rabab and Wang, Xiaofeng and Wang, Song},
|
| 202 |
+
booktitle={Proceedings of the Asian Conference on Computer Vision},
|
| 203 |
+
year={2020}
|
| 204 |
+
}
|
| 205 |
+
|
| 206 |
+
@article{powerlinedataset,
|
| 207 |
+
title={Power image dataset (infrared-ir and visible light-vl)},
|
| 208 |
+
author={Yetgin, {\"O}mer Emre and Gerek, {\"O}mer Nezih and Nezih, {\"O}mer},
|
| 209 |
+
journal={Mendeley Data},
|
| 210 |
+
volume={8},
|
| 211 |
+
year={2017}
|
| 212 |
+
}
|
| 213 |
+
|
| 214 |
+
@inproceedings{cityscapes,
|
| 215 |
+
title={The cityscapes dataset for semantic urban scene understanding},
|
| 216 |
+
author={Cordts, Marius and Omran, Mohamed and Ramos, Sebastian and Rehfeld, Timo and Enzweiler, Markus and Benenson, Rodrigo and Franke, Uwe and Roth, Stefan and Schiele, Bernt},
|
| 217 |
+
booktitle={Proceedings of the IEEE conference on computer vision and pattern recognition},
|
| 218 |
+
pages={3213--3223},
|
| 219 |
+
year={2016}
|
| 220 |
+
}
|
| 221 |
+
|
| 222 |
+
@inproceedings{mapillary,
|
| 223 |
+
title={The mapillary vistas dataset for semantic understanding of street scenes},
|
| 224 |
+
author={Neuhold, Gerhard and Ollmann, Tobias and Rota Bulo, Samuel and Kontschieder, Peter},
|
| 225 |
+
booktitle={Proceedings of the IEEE international conference on computer vision},
|
| 226 |
+
pages={4990--4999},
|
| 227 |
+
year={2017}
|
| 228 |
+
}
|
| 229 |
+
|
| 230 |
+
@inproceedings{coco,
|
| 231 |
+
title={Microsoft coco: Common objects in context},
|
| 232 |
+
author={Lin, Tsung-Yi and Maire, Michael and Belongie, Serge and Hays, James and Perona, Pietro and Ramanan, Deva and Doll{\'a}r, Piotr and Zitnick, C Lawrence},
|
| 233 |
+
booktitle={European conference on computer vision},
|
| 234 |
+
pages={740--755},
|
| 235 |
+
year={2014},
|
| 236 |
+
organization={Springer}
|
| 237 |
+
}
|
| 238 |
+
|
| 239 |
+
@inproceedings{3drecon,
|
| 240 |
+
title={SAIL-VOS 3D: A Synthetic Dataset and Baselines for Object Detection and 3D Mesh Reconstruction from Video Data},
|
| 241 |
+
author={Hu, Yuan-Ting and Wang, Jiahong and Yeh, Raymond A and Schwing, Alexander G},
|
| 242 |
+
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
|
| 243 |
+
pages={1418--1428},
|
| 244 |
+
year={2021}
|
| 245 |
+
}
|
| 246 |
+
|
| 247 |
+
@inproceedings{domainadapt,
|
| 248 |
+
title={Differential treatment for stuff and things: A simple unsupervised domain adaptation method for semantic segmentation},
|
| 249 |
+
author={Wang, Zhonghao and Yu, Mo and Wei, Yunchao and Feris, Rogerio and Xiong, Jinjun and Hwu, Wen-mei and Huang, Thomas S and Shi, Honghui},
|
| 250 |
+
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
|
| 251 |
+
pages={12635--12644},
|
| 252 |
+
year={2020}
|
| 253 |
+
}
|
| 254 |
+
|
| 255 |
+
@article{lsnet,
|
| 256 |
+
title={Ls-net: Fast single-shot line-segment detector},
|
| 257 |
+
author={Nguyen, Van Nhan and Jenssen, Robert and Roverso, Davide},
|
| 258 |
+
journal={arXiv preprint arXiv:1912.09532},
|
| 259 |
+
year={2019}
|
| 260 |
+
}
|
| 261 |
+
|
| 262 |
+
@inproceedings{syntheticwire,
|
| 263 |
+
title={Wire detection using synthetic data and dilated convolutional networks for unmanned aerial vehicles},
|
| 264 |
+
author={Madaan, Ratnesh and Maturana, Daniel and Scherer, Sebastian},
|
| 265 |
+
booktitle={2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)},
|
| 266 |
+
pages={3487--3494},
|
| 267 |
+
year={2017},
|
| 268 |
+
organization={IEEE}
|
| 269 |
+
}
|
| 270 |
+
|
| 271 |
+
@inproceedings{lanedet,
|
| 272 |
+
title={Keep your eyes on the lane: Real-time attention-guided lane detection},
|
| 273 |
+
author={Tabelini, Lucas and Berriel, Rodrigo and Paixao, Thiago M and Badue, Claudine and De Souza, Alberto F and Oliveira-Santos, Thiago},
|
| 274 |
+
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
|
| 275 |
+
pages={294--302},
|
| 276 |
+
year={2021}
|
| 277 |
+
}
|
| 278 |
+
|
| 279 |
+
@article{swiftlane,
|
| 280 |
+
title={SwiftLane: Towards Fast and Efficient Lane Detection},
|
| 281 |
+
author={Jayasinghe, Oshada and Anhettigama, Damith and Hemachandra, Sahan and Kariyawasam, Shenali and Rodrigo, Ranga and Jayasekara, Peshala},
|
| 282 |
+
journal={arXiv preprint arXiv:2110.11779},
|
| 283 |
+
year={2021}
|
| 284 |
+
}
|
| 285 |
+
|
| 286 |
+
@article{structurelane,
|
| 287 |
+
title={Structure Guided Lane Detection},
|
| 288 |
+
author={Su, Jinming and Chen, Chao and Zhang, Ke and Luo, Junfeng and Wei, Xiaoming and Wei, Xiaolin},
|
| 289 |
+
journal={arXiv preprint arXiv:2105.05403},
|
| 290 |
+
year={2021}
|
| 291 |
+
}
|
| 292 |
+
|
| 293 |
+
@inproceedings{linetransformer,
|
| 294 |
+
title={Line segment detection using transformers without edges},
|
| 295 |
+
author={Xu, Yifan and Xu, Weijian and Cheung, David and Tu, Zhuowen},
|
| 296 |
+
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
|
| 297 |
+
pages={4257--4266},
|
| 298 |
+
year={2021}
|
| 299 |
+
}
|
| 300 |
+
|
| 301 |
+
@inproceedings{linegraph,
|
| 302 |
+
title={Ppgnet: Learning point-pair graph for line segment detection},
|
| 303 |
+
author={Zhang, Ziheng and Li, Zhengxin and Bi, Ning and Zheng, Jia and Wang, Jinlei and Huang, Kun and Luo, Weixin and Xu, Yanyu and Gao, Shenghua},
|
| 304 |
+
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
|
| 305 |
+
pages={7105--7114},
|
| 306 |
+
year={2019}
|
| 307 |
+
}
|
| 308 |
+
|
| 309 |
+
@inproceedings{focal,
|
| 310 |
+
title={Focal loss for dense object detection},
|
| 311 |
+
author={Lin, Tsung-Yi and Goyal, Priya and Girshick, Ross and He, Kaiming and Doll{\'a}r, Piotr},
|
| 312 |
+
booktitle={Proceedings of the IEEE international conference on computer vision},
|
| 313 |
+
pages={2980--2988},
|
| 314 |
+
year={2017}
|
| 315 |
+
}
|
| 316 |
+
|
| 317 |
+
@inproceedings{ohem,
|
| 318 |
+
title={Training region-based object detectors with online hard example mining},
|
| 319 |
+
author={Shrivastava, Abhinav and Gupta, Abhinav and Girshick, Ross},
|
| 320 |
+
booktitle={Proceedings of the IEEE conference on computer vision and pattern recognition},
|
| 321 |
+
pages={761--769},
|
| 322 |
+
year={2016}
|
| 323 |
+
}
|
| 324 |
+
|
| 325 |
+
@article{adamw,
|
| 326 |
+
title={Decoupled weight decay regularization},
|
| 327 |
+
author={Loshchilov, Ilya and Hutter, Frank},
|
| 328 |
+
journal={arXiv preprint arXiv:1711.05101},
|
| 329 |
+
year={2017}
|
| 330 |
+
}
|
| 331 |
+
|
| 332 |
+
@article{barnes2009patchmatch,
|
| 333 |
+
title={PatchMatch: A randomized correspondence algorithm for structural image editing},
|
| 334 |
+
author={Barnes, Connelly and Shechtman, Eli and Finkelstein, Adam and Goldman, Dan B},
|
| 335 |
+
journal={ACM Trans. Graph.},
|
| 336 |
+
volume={28},
|
| 337 |
+
number={3},
|
| 338 |
+
pages={24},
|
| 339 |
+
year={2009}
|
| 340 |
+
}
|
| 341 |
+
|
| 342 |
+
@inproceedings{inria,
|
| 343 |
+
title={Can Semantic Labeling Methods Generalize to Any City? The Inria Aerial Image Labeling Benchmark},
|
| 344 |
+
author={Maggiori, Emmanuel and Tarabalka, Yuliya and Charpiat, Guillaume and Alliez, Pierre},
|
| 345 |
+
booktitle={IEEE International Geoscience and Remote Sensing Symposium (IGARSS)},
|
| 346 |
+
year={2017},
|
| 347 |
+
organization={IEEE}
|
| 348 |
+
}
|
| 349 |
+
|
| 350 |
+
@misc{Pixel6as49,
|
| 351 |
+
author = {},
|
| 352 |
+
title = {Pixel 6, a smarter chip for a smarter phone - Google Store},
|
| 353 |
+
howpublished = {\url{https://store.google.com/product/pixel_6?hl=en-US}},
|
| 354 |
+
month = {},
|
| 355 |
+
year = {},
|
| 356 |
+
note = {(Accessed on 11/14/2021)}
|
| 357 |
+
}
|
| 358 |
+
|
| 359 |
+
@inproceedings{isdnet,
|
| 360 |
+
title={ISDNet: Integrating Shallow and Deep Networks for Efficient Ultra-High Resolution Segmentation},
|
| 361 |
+
author={Guo, Shaohua and Liu, Liang and Gan, Zhenye and Wang, Yabiao and Zhang, Wuhao and Wang, Chengjie and Jiang, Guannan and Zhang, Wei and Yi, Ran and Ma, Lizhuang and others},
|
| 362 |
+
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
|
| 363 |
+
pages={4361--4370},
|
| 364 |
+
year={2022}
|
| 365 |
+
}
|
| 366 |
+
@inproceedings{suvorov2022resolution,
|
| 367 |
+
title={Resolution-robust large mask inpainting with fourier convolutions},
|
| 368 |
+
author={Suvorov, Roman and Logacheva, Elizaveta and Mashikhin, Anton and Remizova, Anastasia and Ashukha, Arsenii and Silvestrov, Aleksei and Kong, Naejin and Goka, Harshith and Park, Kiwoong and Lempitsky, Victor},
|
| 369 |
+
booktitle={Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision},
|
| 370 |
+
pages={2149--2159},
|
| 371 |
+
year={2022}
|
| 372 |
+
}
|
| 373 |
+
|
| 374 |
+
@inproceedings{deepglobe,
|
| 375 |
+
title={Deepglobe 2018: A challenge to parse the earth through satellite images},
|
| 376 |
+
author={Demir, Ilke and Koperski, Krzysztof and Lindenbaum, David and Pang, Guan and Huang, Jing and Basu, Saikat and Hughes, Forest and Tuia, Devis and Raskar, Ramesh},
|
| 377 |
+
booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops},
|
| 378 |
+
pages={172--181},
|
| 379 |
+
year={2018}
|
| 380 |
+
}
|
| 381 |
+
|
| 382 |
+
@article{learning_downsample,
|
| 383 |
+
author = {Jin, Chen
|
| 384 |
+
and Tanno, Ryutaro
|
| 385 |
+
and Mertzanidou, Thomy
|
| 386 |
+
and Panagiotaki, Eleftheria
|
| 387 |
+
and Alexander, Daniel C},
|
| 388 |
+
title = {{Learning to Downsample for Segmentation of Ultra-High Resolution Images}},
|
| 389 |
+
journal = {arXiv preprint arXiv:2109.11071},
|
| 390 |
+
year = {2021}
|
| 391 |
+
}
|
| 392 |
+
|
| 393 |
+
@inproceedings{stdc,
|
| 394 |
+
title={Rethinking BiSeNet for real-time semantic segmentation},
|
| 395 |
+
author={Fan, Mingyuan and Lai, Shenqi and Huang, Junshi and Wei, Xiaoming and Chai, Zhenhua and Luo, Junfeng and Wei, Xiaolin},
|
| 396 |
+
booktitle={Proceedings of the IEEE/CVF conference on computer vision and pattern recognition},
|
| 397 |
+
pages={9716--9725},
|
| 398 |
+
year={2021}
|
| 399 |
+
}
|
| 400 |
+
|
| 401 |
+
@article{zheng2022cm,
|
| 402 |
+
title={CM-GAN: Image Inpainting with Cascaded Modulation GAN and Object-Aware Training},
|
| 403 |
+
author={Zheng, Haitian and Lin, Zhe and Lu, Jingwan and Cohen, Scott and Shechtman, Eli and Barnes, Connelly and Zhang, Jianming and Xu, Ning and Amirghodsi, Sohrab and Luo, Jiebo},
|
| 404 |
+
journal={arXiv preprint arXiv:2203.11947},
|
| 405 |
+
year={2022}
|
| 406 |
+
}
|
| 407 |
+
|
| 408 |
+
@article{jain2022keys,
|
| 409 |
+
title={Keys to Better Image Inpainting: Structure and Texture Go Hand in Hand},
|
| 410 |
+
author={Jain, Jitesh and Zhou, Yuqian and Yu, Ning and Shi, Humphrey},
|
| 411 |
+
journal={arXiv preprint arXiv:2208.03382},
|
| 412 |
+
year={2022}
|
| 413 |
+
}
|
| 414 |
+
|
| 415 |
+
@inproceedings{rombach2022high,
|
| 416 |
+
title={High-resolution image synthesis with latent diffusion models},
|
| 417 |
+
author={Rombach, Robin and Blattmann, Andreas and Lorenz, Dominik and Esser, Patrick and Ommer, Bj{\"o}rn},
|
| 418 |
+
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
|
| 419 |
+
pages={10684--10695},
|
| 420 |
+
year={2022}
|
| 421 |
+
}
|
| 422 |
+
|
| 423 |
+
@inproceedings{yu2019free,
|
| 424 |
+
title={Free-form image inpainting with gated convolution},
|
| 425 |
+
author={Yu, Jiahui and Lin, Zhe and Yang, Jimei and Shen, Xiaohui and Lu, Xin and Huang, Thomas S},
|
| 426 |
+
booktitle={Proceedings of the IEEE/CVF international conference on computer vision},
|
| 427 |
+
pages={4471--4480},
|
| 428 |
+
year={2019}
|
| 429 |
+
}
|
| 430 |
+
|
| 431 |
+
@article{zhou2017places,
|
| 432 |
+
title={Places: A 10 million Image Database for Scene Recognition},
|
| 433 |
+
author={Zhou, Bolei and Lapedriza, Agata and Khosla, Aditya and Oliva, Aude and Torralba, Antonio},
|
| 434 |
+
journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
|
| 435 |
+
year={2017},
|
| 436 |
+
publisher={IEEE}
|
| 437 |
+
}
|
| 438 |
+
|
| 439 |
+
@inproceedings{karras2020analyzing,
|
| 440 |
+
title={Analyzing and improving the image quality of stylegan},
|
| 441 |
+
author={Karras, Tero and Laine, Samuli and Aittala, Miika and Hellsten, Janne and Lehtinen, Jaakko and Aila, Timo},
|
| 442 |
+
booktitle={Proceedings of the IEEE/CVF conference on computer vision and pattern recognition},
|
| 443 |
+
pages={8110--8119},
|
| 444 |
+
year={2020}
|
| 445 |
+
}
|
| 446 |
+
|
| 447 |
+
@article{wexler2007space,
|
| 448 |
+
title={Space-time completion of video},
|
| 449 |
+
author={Wexler, Yonatan and Shechtman, Eli and Irani, Michal},
|
| 450 |
+
journal={IEEE Transactions on pattern analysis and machine intelligence},
|
| 451 |
+
volume={29},
|
| 452 |
+
number={3},
|
| 453 |
+
pages={463--476},
|
| 454 |
+
year={2007},
|
| 455 |
+
publisher={IEEE}
|
| 456 |
+
}
|
| 457 |
+
|
| 458 |
+
@article{darabi2012image,
|
| 459 |
+
title={Image melding: Combining inconsistent images using patch-based synthesis},
|
| 460 |
+
author={Darabi, Soheil and Shechtman, Eli and Barnes, Connelly and Goldman, Dan B and Sen, Pradeep},
|
| 461 |
+
journal={ACM Transactions on graphics (TOG)},
|
| 462 |
+
volume={31},
|
| 463 |
+
number={4},
|
| 464 |
+
pages={1--10},
|
| 465 |
+
year={2012},
|
| 466 |
+
publisher={ACM New York, NY, USA}
|
| 467 |
+
}
|
| 468 |
+
|
| 469 |
+
@inproceedings{kaspar2015self,
|
| 470 |
+
title={Self tuning texture optimization},
|
| 471 |
+
author={Kaspar, Alexandre and Neubert, Boris and Lischinski, Dani and Pauly, Mark and Kopf, Johannes},
|
| 472 |
+
booktitle={Computer Graphics Forum},
|
| 473 |
+
volume={34},
|
| 474 |
+
number={2},
|
| 475 |
+
pages={349--359},
|
| 476 |
+
year={2015},
|
| 477 |
+
organization={Wiley Online Library}
|
| 478 |
+
}
|
| 479 |
+
|
| 480 |
+
@inproceedings{contextencoder,
|
| 481 |
+
title={Context encoders: Feature learning by inpainting},
|
| 482 |
+
author={Pathak, Deepak and Krahenbuhl, Philipp and Donahue, Jeff and Darrell, Trevor and Efros, Alexei A},
|
| 483 |
+
booktitle={Proceedings of the IEEE conference on computer vision and pattern recognition},
|
| 484 |
+
pages={2536--2544},
|
| 485 |
+
year={2016}
|
| 486 |
+
}
|
| 487 |
+
|
| 488 |
+
@article{globallocal,
|
| 489 |
+
title={Globally and locally consistent image completion},
|
| 490 |
+
author={Iizuka, Satoshi and Simo-Serra, Edgar and Ishikawa, Hiroshi},
|
| 491 |
+
journal={ACM Transactions on Graphics (ToG)},
|
| 492 |
+
volume={36},
|
| 493 |
+
number={4},
|
| 494 |
+
pages={1--14},
|
| 495 |
+
year={2017},
|
| 496 |
+
publisher={ACM New York, NY, USA}
|
| 497 |
+
}
|
| 498 |
+
|
| 499 |
+
@inproceedings{contextual,
|
| 500 |
+
title={Generative image inpainting with contextual attention},
|
| 501 |
+
author={Yu, Jiahui and Lin, Zhe and Yang, Jimei and Shen, Xiaohui and Lu, Xin and Huang, Thomas S},
|
| 502 |
+
booktitle={Proceedings of the IEEE conference on computer vision and pattern recognition},
|
| 503 |
+
pages={5505--5514},
|
| 504 |
+
year={2018}
|
| 505 |
+
}
|
| 506 |
+
|
| 507 |
+
@inproceedings{partialconv,
|
| 508 |
+
title={Image inpainting for irregular holes using partial convolutions},
|
| 509 |
+
author={Liu, Guilin and Reda, Fitsum A and Shih, Kevin J and Wang, Ting-Chun and Tao, Andrew and Catanzaro, Bryan},
|
| 510 |
+
booktitle={Proceedings of the European conference on computer vision (ECCV)},
|
| 511 |
+
pages={85--100},
|
| 512 |
+
year={2018}
|
| 513 |
+
}
|
| 514 |
+
@article{comodgan,
|
| 515 |
+
title={Large scale image completion via co-modulated generative adversarial networks},
|
| 516 |
+
author={Zhao, Shengyu and Cui, Jonathan and Sheng, Yilun and Dong, Yue and Liang, Xiao and Chang, Eric I and Xu, Yan},
|
| 517 |
+
journal={arXiv preprint arXiv:2103.10428},
|
| 518 |
+
year={2021}
|
| 519 |
+
}
|
| 520 |
+
|
| 521 |
+
@article{dalle,
|
| 522 |
+
title={Hierarchical text-conditional image generation with clip latents},
|
| 523 |
+
author={Ramesh, Aditya and Dhariwal, Prafulla and Nichol, Alex and Chu, Casey and Chen, Mark},
|
| 524 |
+
journal={arXiv preprint arXiv:2204.06125},
|
| 525 |
+
year={2022}
|
| 526 |
+
}
|
| 527 |
+
|
| 528 |
+
@inproceedings{zeng2020high,
|
| 529 |
+
title={High-resolution image inpainting with iterative confidence feedback and guided upsampling},
|
| 530 |
+
author={Zeng, Yu and Lin, Zhe and Yang, Jimei and Zhang, Jianming and Shechtman, Eli and Lu, Huchuan},
|
| 531 |
+
booktitle={European conference on computer vision},
|
| 532 |
+
pages={1--17},
|
| 533 |
+
year={2020},
|
| 534 |
+
organization={Springer}
|
| 535 |
+
}
|
| 536 |
+
|
| 537 |
+
|
| 538 |
+
@inproceedings{hifill,
|
| 539 |
+
title={Contextual Residual Aggregation for Ultra High-Resolution Image Inpainting},
|
| 540 |
+
author={Yi, Zili and Tang, Qiang and Azizi, Shekoofeh and Jang, Daesik and Xu, Zhan},
|
| 541 |
+
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
|
| 542 |
+
pages={7508--7517},
|
| 543 |
+
year={2020}
|
| 544 |
+
}
|
| 545 |
+
@inproceedings{supercaf,
|
| 546 |
+
title={Inpainting at Modern Camera Resolution by Guided PatchMatch with Auto-curation},
|
| 547 |
+
author={Zhang, Lingzhi and Barnes, Connelly and Wampler, Kevin and Amirghodsi, Sohrab and Shechtman, Eli and Lin, Zhe and Shi, Jianbo},
|
| 548 |
+
booktitle={European Conference on Computer Vision},
|
| 549 |
+
pages={51--67},
|
| 550 |
+
year={2022},
|
| 551 |
+
organization={Springer}
|
| 552 |
+
}
|
| 553 |
+
|
| 554 |
+
@article{hassani2022neighborhood,
|
| 555 |
+
title = {Neighborhood Attention Transformer},
|
| 556 |
+
author = {Ali Hassani and Steven Walton and Jiachen Li and Shen Li and Humphrey Shi},
|
| 557 |
+
year = 2022,
|
| 558 |
+
url = {https://arxiv.org/abs/2204.07143},
|
| 559 |
+
eprint = {2204.07143},
|
| 560 |
+
archiveprefix = {arXiv},
|
| 561 |
+
primaryclass = {cs.CV}
|
| 562 |
+
}
|
| 563 |
+
@article{hassani2022dilated,
|
| 564 |
+
title = {Dilated Neighborhood Attention Transformer},
|
| 565 |
+
author = {Ali Hassani and Humphrey Shi},
|
| 566 |
+
year = 2022,
|
| 567 |
+
url = {https://arxiv.org/abs/2209.15001},
|
| 568 |
+
eprint = {2209.15001},
|
| 569 |
+
archiveprefix = {arXiv},
|
| 570 |
+
primaryclass = {cs.CV}
|
| 571 |
+
}
|
| 572 |
+
|
| 573 |
+
@article{jain2021semask,
|
| 574 |
+
title={SeMask: Semantically Masking Transformer Backbones for Effective Semantic Segmentation},
|
| 575 |
+
author={Jitesh Jain and Anukriti Singh and Nikita Orlov and Zilong Huang and Jiachen Li and Steven Walton and Humphrey Shi},
|
| 576 |
+
journal={arXiv},
|
| 577 |
+
year={2021}
|
| 578 |
+
}
|
| 579 |
+
|
| 580 |
+
@inproceedings{jain2022oneformer,
|
| 581 |
+
title={{OneFormer: One Transformer to Rule Universal Image Segmentation}},
|
| 582 |
+
author={Jitesh Jain and Jiachen Li and MangTik Chiu and Ali Hassani and Nikita Orlov and Humphrey Shi},
|
| 583 |
+
journal={CVPR},
|
| 584 |
+
year={2023}
|
| 585 |
+
}
|
| 586 |
+
|
| 587 |
+
@inproceedings{xu2022image,
|
| 588 |
+
title={Image Completion with Heterogeneously Filtered Spectral Hints},
|
| 589 |
+
author={Xingqian Xu and Shant Navasardyan and Vahram Tadevosyan and Andranik Sargsyan and Yadong Mu and Humphrey Shi},
|
| 590 |
+
booktitle={WACV},
|
| 591 |
+
year={2023}
|
| 592 |
+
}
|
paper-tex/figure_tex/additional_visualizations.tex
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\begin{figure*}[t!]
|
| 3 |
+
\setlength{\abovecaptionskip}{1mm}
|
| 4 |
+
\centering
|
| 5 |
+
\includegraphics[width=\textwidth]{figures/additional_visualizations.pdf}
|
| 6 |
+
\caption{\textbf{Segmentation and inpainting visualizations}. Our model can handle several challenging scenes, including strongly backlit (top row), background with complex texture (2nd row), low light (3rd row), and barely visible wires (4th row)}
|
| 7 |
+
\label{fig:additional_visualizations}
|
| 8 |
+
\end{figure*}
|
paper-tex/figure_tex/annotation.tex
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\begin{figure}[t!]
|
| 3 |
+
\centering
|
| 4 |
+
\captionsetup{type=figure}
|
| 5 |
+
\includegraphics[width=\linewidth]{figures/annotation.pdf}\\
|
| 6 |
+
\vspace{-1mm}
|
| 7 |
+
\captionof{figure}{\textbf{Wire Annotation Example.} An example wire annotation in our dataset. Our annotation (B) is accurate in different wire thicknesses (\textcolor{red}{red}), variations in wire shapes (\textcolor{orange}{orange}) and accurate wire occlusions (\textcolor{yellow}{yellow}).}
|
| 8 |
+
% \textcolor{red}{This image is unchanged from last year, do we want to say this is one of the collected images rather than saying it's in the benchmark dataset?}
|
| 9 |
+
\vspace{-4mm}
|
| 10 |
+
\label{fig:annotation}
|
| 11 |
+
\end{figure}
|
paper-tex/figure_tex/comparison.tex
ADDED
|
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\begin{figure*}[t!]
|
| 3 |
+
\centering
|
| 4 |
+
% \vspace{1mm}
|
| 5 |
+
\includegraphics[width=1.0\linewidth]{figures/wire-qualitative.pdf}
|
| 6 |
+
\vspace{-8mm}
|
| 7 |
+
\caption{
|
| 8 |
+
% Qualitative comparison between our method and baseline networks on a 32 Megapixel image. (a), (b): Whole image inference. (c), (d): Sliding window inference. (e), (f): Our two-stage model. Our model predicts much tighter segmentation masks without artifacts such as aliasing, while avoiding false positives caused by only seeing local regions. Each R/G/B bounding box in (c),(d),(e) represents a 1024$\times$1024 sliding-window used by the fine module for refinement.
|
| 9 |
+
\textbf{Qualitative comparison of several semantic segmentation models.} %Our model predicts much tighter and consistent segmentation masks, while avoiding false positives caused by only seeing local regions. Note that in the third row, our two-stage model successfully suppresses false positives that would otherwise be misclassified without global context.
|
| 10 |
+
A common object semantic segmentation model (DeepLabv3+) either fails to find thin wires or overpredicts due to lack of global context. On the other hand, CascadePSP and MagNet, being refinement-based models, cannot work well on wires when the predictions are inaccurate or missing. While ISDNet can capture many thin wires regions, it cannot produce a high-quality prediction. In contrast, our model is able to both capture accurate wire regions and produce fine wire masks, and maintain low inference time.
|
| 11 |
+
\vspace{-5mm}
|
| 12 |
+
}
|
| 13 |
+
\label{fig:visual}
|
| 14 |
+
\end{figure*}
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
\begin{figure*}[t!]
|
| 18 |
+
\centering
|
| 19 |
+
\vspace{1mm}
|
| 20 |
+
\includegraphics[width=1.0\linewidth]{figures/wire-ablation.pdf}
|
| 21 |
+
\vspace{-7mm}
|
| 22 |
+
\caption{
|
| 23 |
+
\textbf{Qualitative comparison of our model components.} MinMax enhances wire image features when they are too subtle to see in RGB, while MaxPool encourages aggressive predictions in the coarse branch. Both components enable the model to pick up more regions for the final wire mask prediction.
|
| 24 |
+
}
|
| 25 |
+
% \vspace{-2mm}
|
| 26 |
+
\label{fig:visual}
|
| 27 |
+
\end{figure*}
|
paper-tex/figure_tex/failure.tex
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\begin{figure}[h!]
|
| 3 |
+
\centering
|
| 4 |
+
\captionsetup{type=figure}
|
| 5 |
+
\includegraphics[width=1.\linewidth]{figures/failure.pdf}
|
| 6 |
+
\vspace{-6mm}
|
| 7 |
+
\captionof{figure}{\textbf{Failure cases}. Our model can fail for cases where wires are heavily blended with the background (upper row), or on thick wires with complex patterns that are rarely seen.
|
| 8 |
+
\vspace{-5mm}
|
| 9 |
+
}
|
| 10 |
+
\label{fig:failure}
|
| 11 |
+
\end{figure}
|
paper-tex/figure_tex/inpaint_compare.tex
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
|
| 3 |
+
\begin{figure}[h!]
|
| 4 |
+
\centering
|
| 5 |
+
\captionsetup{type=figure}
|
| 6 |
+
\includegraphics[width=1.\linewidth]{figures/inpainting_result.pdf}
|
| 7 |
+
\vspace{-6mm}
|
| 8 |
+
\captionof{figure}{\textbf{Inpainting Comparison}. Our model performs well on complicated structure completion and color consistency, especially on building facades and sky regions containing plain and uniform color.
|
| 9 |
+
\vspace{-3mm}
|
| 10 |
+
}
|
| 11 |
+
\label{fig:wire_inp}
|
| 12 |
+
\end{figure}
|
paper-tex/figure_tex/motivation.tex
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\begin{figure}[t!]
|
| 3 |
+
\centering
|
| 4 |
+
\includegraphics[width=1.0\linewidth]{figures/wireishard.pdf}
|
| 5 |
+
% \includegraphics[width=1.0\linewidth]{figures/wiresarehard2.pdf}
|
| 6 |
+
\caption{\textbf{Challenges of wire segmentation.} Wires have a diverse set of appearances. Challenges include but are not limited to (a) structural complexity, (b) visibility and thickness, (c) partial occlusion by other objects, (d) camera aberration artifacts, and variations in (e) object attachment, (f) color, (g) width and (h) shape.
|
| 7 |
+
% \zwei{this needs to be correspondent to the attributes you mentioned}
|
| 8 |
+
}
|
| 9 |
+
\vspace{-5.5mm}
|
| 10 |
+
\label{fig:motivation}
|
| 11 |
+
\end{figure}
|
paper-tex/figure_tex/new_failure_cases.tex
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\begin{figure}[h!]
|
| 3 |
+
\setlength{\abovecaptionskip}{1mm}
|
| 4 |
+
\centering
|
| 5 |
+
\includegraphics[width=0.9\linewidth]{figures/new_failures.pdf}
|
| 6 |
+
\caption{\textbf{Failure cases}. In some challenging cases, our model fails to predict accurate masks. Zoom in to see detailed wire masks in ground truth and prediction.}
|
| 7 |
+
\label{fig:new_failures}
|
| 8 |
+
\vspace{-10mm}
|
| 9 |
+
\end{figure}
|
paper-tex/figure_tex/new_panorama.tex
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\begin{figure*}[t!]
|
| 3 |
+
\centering
|
| 4 |
+
\includegraphics[width=1.0\textwidth]{figures/new_panorama.pdf}
|
| 5 |
+
\caption{\textbf{Segmentation and inpainting result for a panoramic image.} Our model is scalable to very large images with very thin wires.}
|
| 6 |
+
\label{fig:new_panorama}
|
| 7 |
+
\vspace{-3mm}
|
| 8 |
+
\end{figure*}
|
paper-tex/figure_tex/overpredict.tex
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\begin{figure}[h!]
|
| 3 |
+
\centering
|
| 4 |
+
\begin{tabular}{@{}c@{\hspace{1mm}}c@{\hspace{1mm}}c@{}}
|
| 5 |
+
\captionsetup{type=figure}
|
| 6 |
+
\includegraphics[width=0.9\linewidth]{figures/overpredict/02709_global.jpg} \\
|
| 7 |
+
(a) Fine module output with global logit map \\
|
| 8 |
+
and binary location map. \\
|
| 9 |
+
\includegraphics[width=0.9\linewidth]{figures/overpredict/02709_local.jpg} \\
|
| 10 |
+
(b) Fine module output with only local logit map. \\
|
| 11 |
+
\end{tabular}
|
| 12 |
+
\captionof{figure}{(Image to be updated)}
|
| 13 |
+
\label{fig:overpredict}
|
| 14 |
+
\end{figure}
|
paper-tex/figure_tex/panorama.tex
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\begin{figure*}[t!]
|
| 3 |
+
\centering
|
| 4 |
+
\includegraphics[width=1.0\textwidth]{figures/panorama.pdf}
|
| 5 |
+
\vspace{-7mm}
|
| 6 |
+
\caption{\textbf{Segmentation result for a panoramic image.} Our two-stage model leverages the sparsity of wires in natural images, and efficiently generalizes to ultra-high resolution images such as this panorama of $10$K by $2$K resolution. Note that our method produces high quality wire segmentation that covers wires that are almost invisible.}
|
| 7 |
+
\vspace{-4mm}
|
| 8 |
+
\label{fig:panorama}
|
| 9 |
+
\end{figure*}
|
paper-tex/figure_tex/pipeline.tex
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\begin{figure*}[hbt!]
|
| 3 |
+
\centering
|
| 4 |
+
\includegraphics[width=1.0\textwidth]{figures/wire-pipeline.png}
|
| 5 |
+
\vspace{-2mm}
|
| 6 |
+
\caption{\textbf{Our wire removal system}. A system overview of our wire segmentation and removal for high resolution images. Input is concatenated with min- and max-filtered luminance channels. The downsampled input is fed into the coarse module to obtain the global probability. In the local stage, original-resolution patches are concatenated with the global probability map to obtain the local logit map. After a segmentation mask is predicted, we adopt LaMa architecture and use a tile-based approach to achieve wire removal. See Section~\ref{sec:segmentation},~\ref{sec:inpainting} for details.}
|
| 7 |
+
\vspace{-1mm}
|
| 8 |
+
\label{fig:pipeline}
|
| 9 |
+
\end{figure*}
|
paper-tex/figure_tex/pixel6.tex
ADDED
|
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\begin{figure}[h!]
|
| 3 |
+
\setlength{\abovecaptionskip}{1mm}
|
| 4 |
+
\centering
|
| 5 |
+
\captionsetup{type=figure}
|
| 6 |
+
\includegraphics[width=0.43\textwidth]{figures/wire-pixel6.pdf}
|
| 7 |
+
\vspace{-1mm}
|
| 8 |
+
\captionof{figure}{\textbf{Comparison with Pixel 6}. Our model can pick up hardly visible wires that even in complicated backgrounds}
|
| 9 |
+
\label{fig:pixel6}
|
| 10 |
+
\end{figure}
|
paper-tex/figure_tex/teaser.tex
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\twocolumn[{%
|
| 3 |
+
\renewcommand\twocolumn[1][]{#1}%
|
| 4 |
+
\begin{center}
|
| 5 |
+
\maketitle
|
| 6 |
+
\centering
|
| 7 |
+
\begin{tabular}{@{}c@{\hspace{1mm}}c@{\hspace{1mm}}c@{}}
|
| 8 |
+
\captionsetup{type=figure}
|
| 9 |
+
\includegraphics[width=0.5\linewidth]{figures/teaser/Picture2.jpg} &
|
| 10 |
+
\includegraphics[width=0.5\linewidth]{figures/teaser/Picture4.jpg} \\
|
| 11 |
+
Input Image & Our Wire Segmentation Result \\
|
| 12 |
+
% \includegraphics[width=0.33\linewidth]{figures/teaser/1_caf.png} \\
|
| 13 |
+
\end{tabular}
|
| 14 |
+
\captionof{figure}{\textbf{Our wire-like object segmentation result}. The input image (left) is a 12 Megapixel image taken by a smartphone. Our wire-like object segmentation shows high quality result (right) despite the complexity and variation of wires in the input image. (All figures in this paper are best viewed in full resolution on a big screen.)}
|
| 15 |
+
\end{center}%
|
| 16 |
+
}]
|
paper-tex/figures/additional_visualizations.pdf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6c54751212b9a552233f3108ef66ad62464edd4777f23d59e41ecbeae2f09fc7
|
| 3 |
+
size 826719
|
paper-tex/figures/annotation.pdf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:06c9b1a5acf8058b220e41e16a7fb1deabbcd27a2066e715cad7386bec976035
|
| 3 |
+
size 201747
|
paper-tex/figures/inpainting_result.pdf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8cfd220fae94816922173c8ab3d46c7b46601f31059f9490ca18e228502094fa
|
| 3 |
+
size 4586003
|
paper-tex/figures/new_failures.pdf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5b27c6a579f3029e483f8ee6ebeae281f330f7567fa63271fcaf8862e5971602
|
| 3 |
+
size 121316
|
paper-tex/figures/new_panorama.pdf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fe64a9b8f05d7fdd0eb5e0b4716d3e0c50009b75d7f960b9221cf71306362fc2
|
| 3 |
+
size 865378
|
paper-tex/figures/panorama.pdf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:571daeb6fe856e769e3b0911e63207dca2ab7ec7800a9b9e071ecee60e0df9a2
|
| 3 |
+
size 2475508
|
paper-tex/figures/pixel6.pdf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:65360277495b27fc7bba97c7c23b2abe4f3bc56dfe7edef18b70710103fc3532
|
| 3 |
+
size 2317563
|
paper-tex/figures/qualitative.pdf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3da1671ec43dbc8734c06f15c573b39b34eef7ba71da3e3deef8293f13181149
|
| 3 |
+
size 1232647
|
paper-tex/figures/teaser2.pdf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d0479a842e86f66c4b3a2963fe9c9f23885fa6a6c6dc7a8e1bbe056e054d2c59
|
| 3 |
+
size 8619642
|
paper-tex/figures/wire-ablation.pdf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d6b302521df024fa15cd61470eb367507e8fd1f97adb6cee9b18f8882596c81a
|
| 3 |
+
size 353551
|
paper-tex/figures/wire-pipeline.png
ADDED
|
Git LFS Details
|
paper-tex/figures/wire-pixel6.pdf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:094445913469d368179893224d55e60ca5c4ad21d8506aa495e0b5734c1abe6b
|
| 3 |
+
size 10657467
|
paper-tex/figures/wire-qualitative.pdf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5484ce337ed151049fd89f9a05f190f22cda15c97ee4d3911312752c80f5f4a1
|
| 3 |
+
size 588630
|
paper-tex/figures/wireishard.pdf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b8153a4753f1c270f06855f724a2d0ebe8e026cf4a035928ae5cc666f5fef6f
|
| 3 |
+
size 86567
|
paper-tex/ieee_fullname.bst
ADDED
|
@@ -0,0 +1,1135 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
% This is a modification to the normal ieee.bst used by CVPR to render
|
| 2 |
+
% first names in the bibliography as "Firstname Lastname" rather than
|
| 3 |
+
% "F. Lastname".
|
| 4 |
+
%
|
| 5 |
+
% Jonathan T. Barron, 12/5/2018, [email protected]
|
| 6 |
+
|
| 7 |
+
% ---------------------------------------------------------------
|
| 8 |
+
%
|
| 9 |
+
% ieee.bst,v 1.0 2002/04/16
|
| 10 |
+
%
|
| 11 |
+
% by Glenn Paulley ([email protected])
|
| 12 |
+
%
|
| 13 |
+
% Modified from latex8.bst 1995/09/15 15:13:49 ienne Exp $
|
| 14 |
+
%
|
| 15 |
+
% by [email protected]
|
| 16 |
+
%
|
| 17 |
+
%
|
| 18 |
+
% ---------------------------------------------------------------
|
| 19 |
+
%
|
| 20 |
+
% no guarantee is given that the format corresponds perfectly to
|
| 21 |
+
% IEEE 8.5" x 11" Proceedings, but most features should be ok.
|
| 22 |
+
%
|
| 23 |
+
% ---------------------------------------------------------------
|
| 24 |
+
%
|
| 25 |
+
% `ieee' from BibTeX standard bibliography style `abbrv'
|
| 26 |
+
% version 0.99a for BibTeX versions 0.99a or later, LaTeX version 2.09.
|
| 27 |
+
% Copyright (C) 1985, all rights reserved.
|
| 28 |
+
% Copying of this file is authorized only if either
|
| 29 |
+
% (1) you make absolutely no changes to your copy, including name, or
|
| 30 |
+
% (2) if you do make changes, you name it something other than
|
| 31 |
+
% btxbst.doc, plain.bst, unsrt.bst, alpha.bst, and abbrv.bst.
|
| 32 |
+
% This restriction helps ensure that all standard styles are identical.
|
| 33 |
+
% The file btxbst.doc has the documentation for this style.
|
| 34 |
+
|
| 35 |
+
ENTRY
|
| 36 |
+
{ address
|
| 37 |
+
author
|
| 38 |
+
booktitle
|
| 39 |
+
chapter
|
| 40 |
+
edition
|
| 41 |
+
editor
|
| 42 |
+
howpublished
|
| 43 |
+
institution
|
| 44 |
+
journal
|
| 45 |
+
key
|
| 46 |
+
month
|
| 47 |
+
note
|
| 48 |
+
number
|
| 49 |
+
organization
|
| 50 |
+
pages
|
| 51 |
+
publisher
|
| 52 |
+
school
|
| 53 |
+
series
|
| 54 |
+
title
|
| 55 |
+
type
|
| 56 |
+
volume
|
| 57 |
+
year
|
| 58 |
+
}
|
| 59 |
+
{}
|
| 60 |
+
{ label }
|
| 61 |
+
|
| 62 |
+
INTEGERS { output.state before.all mid.sentence after.sentence after.block }
|
| 63 |
+
|
| 64 |
+
FUNCTION {init.state.consts}
|
| 65 |
+
{ #0 'before.all :=
|
| 66 |
+
#1 'mid.sentence :=
|
| 67 |
+
#2 'after.sentence :=
|
| 68 |
+
#3 'after.block :=
|
| 69 |
+
}
|
| 70 |
+
|
| 71 |
+
STRINGS { s t }
|
| 72 |
+
|
| 73 |
+
FUNCTION {output.nonnull}
|
| 74 |
+
{ 's :=
|
| 75 |
+
output.state mid.sentence =
|
| 76 |
+
{ ", " * write$ }
|
| 77 |
+
{ output.state after.block =
|
| 78 |
+
{ add.period$ write$
|
| 79 |
+
newline$
|
| 80 |
+
"\newblock " write$
|
| 81 |
+
}
|
| 82 |
+
{ output.state before.all =
|
| 83 |
+
'write$
|
| 84 |
+
{ add.period$ " " * write$ }
|
| 85 |
+
if$
|
| 86 |
+
}
|
| 87 |
+
if$
|
| 88 |
+
mid.sentence 'output.state :=
|
| 89 |
+
}
|
| 90 |
+
if$
|
| 91 |
+
s
|
| 92 |
+
}
|
| 93 |
+
|
| 94 |
+
FUNCTION {output}
|
| 95 |
+
{ duplicate$ empty$
|
| 96 |
+
'pop$
|
| 97 |
+
'output.nonnull
|
| 98 |
+
if$
|
| 99 |
+
}
|
| 100 |
+
|
| 101 |
+
FUNCTION {output.check}
|
| 102 |
+
{ 't :=
|
| 103 |
+
duplicate$ empty$
|
| 104 |
+
{ pop$ "empty " t * " in " * cite$ * warning$ }
|
| 105 |
+
'output.nonnull
|
| 106 |
+
if$
|
| 107 |
+
}
|
| 108 |
+
|
| 109 |
+
FUNCTION {output.bibitem}
|
| 110 |
+
{ newline$
|
| 111 |
+
"\bibitem{" write$
|
| 112 |
+
cite$ write$
|
| 113 |
+
"}" write$
|
| 114 |
+
newline$
|
| 115 |
+
""
|
| 116 |
+
before.all 'output.state :=
|
| 117 |
+
}
|
| 118 |
+
|
| 119 |
+
FUNCTION {fin.entry}
|
| 120 |
+
{ add.period$
|
| 121 |
+
write$
|
| 122 |
+
newline$
|
| 123 |
+
}
|
| 124 |
+
|
| 125 |
+
FUNCTION {new.block}
|
| 126 |
+
{ output.state before.all =
|
| 127 |
+
'skip$
|
| 128 |
+
{ after.block 'output.state := }
|
| 129 |
+
if$
|
| 130 |
+
}
|
| 131 |
+
|
| 132 |
+
FUNCTION {new.sentence}
|
| 133 |
+
{ output.state after.block =
|
| 134 |
+
'skip$
|
| 135 |
+
{ output.state before.all =
|
| 136 |
+
'skip$
|
| 137 |
+
{ after.sentence 'output.state := }
|
| 138 |
+
if$
|
| 139 |
+
}
|
| 140 |
+
if$
|
| 141 |
+
}
|
| 142 |
+
|
| 143 |
+
FUNCTION {not}
|
| 144 |
+
{ { #0 }
|
| 145 |
+
{ #1 }
|
| 146 |
+
if$
|
| 147 |
+
}
|
| 148 |
+
|
| 149 |
+
FUNCTION {and}
|
| 150 |
+
{ 'skip$
|
| 151 |
+
{ pop$ #0 }
|
| 152 |
+
if$
|
| 153 |
+
}
|
| 154 |
+
|
| 155 |
+
FUNCTION {or}
|
| 156 |
+
{ { pop$ #1 }
|
| 157 |
+
'skip$
|
| 158 |
+
if$
|
| 159 |
+
}
|
| 160 |
+
|
| 161 |
+
FUNCTION {new.block.checka}
|
| 162 |
+
{ empty$
|
| 163 |
+
'skip$
|
| 164 |
+
'new.block
|
| 165 |
+
if$
|
| 166 |
+
}
|
| 167 |
+
|
| 168 |
+
FUNCTION {new.block.checkb}
|
| 169 |
+
{ empty$
|
| 170 |
+
swap$ empty$
|
| 171 |
+
and
|
| 172 |
+
'skip$
|
| 173 |
+
'new.block
|
| 174 |
+
if$
|
| 175 |
+
}
|
| 176 |
+
|
| 177 |
+
FUNCTION {new.sentence.checka}
|
| 178 |
+
{ empty$
|
| 179 |
+
'skip$
|
| 180 |
+
'new.sentence
|
| 181 |
+
if$
|
| 182 |
+
}
|
| 183 |
+
|
| 184 |
+
FUNCTION {new.sentence.checkb}
|
| 185 |
+
{ empty$
|
| 186 |
+
swap$ empty$
|
| 187 |
+
and
|
| 188 |
+
'skip$
|
| 189 |
+
'new.sentence
|
| 190 |
+
if$
|
| 191 |
+
}
|
| 192 |
+
|
| 193 |
+
FUNCTION {field.or.null}
|
| 194 |
+
{ duplicate$ empty$
|
| 195 |
+
{ pop$ "" }
|
| 196 |
+
'skip$
|
| 197 |
+
if$
|
| 198 |
+
}
|
| 199 |
+
|
| 200 |
+
FUNCTION {emphasize}
|
| 201 |
+
{ duplicate$ empty$
|
| 202 |
+
{ pop$ "" }
|
| 203 |
+
{ "{\em " swap$ * "}" * }
|
| 204 |
+
if$
|
| 205 |
+
}
|
| 206 |
+
|
| 207 |
+
INTEGERS { nameptr namesleft numnames }
|
| 208 |
+
|
| 209 |
+
FUNCTION {format.names}
|
| 210 |
+
{ 's :=
|
| 211 |
+
#1 'nameptr :=
|
| 212 |
+
s num.names$ 'numnames :=
|
| 213 |
+
numnames 'namesleft :=
|
| 214 |
+
{ namesleft #0 > }
|
| 215 |
+
% Formerly { s nameptr "{f.~}{vv~}{ll}{, jj}" format.name$ 't :=
|
| 216 |
+
{ s nameptr "{ff }{vv }{ll}{, jj}" format.name$ 't :=
|
| 217 |
+
nameptr #1 >
|
| 218 |
+
{ namesleft #1 >
|
| 219 |
+
{ ", " * t * }
|
| 220 |
+
{ numnames #2 >
|
| 221 |
+
{ "," * }
|
| 222 |
+
'skip$
|
| 223 |
+
if$
|
| 224 |
+
t "others" =
|
| 225 |
+
{ " et~al." * }
|
| 226 |
+
{ " and " * t * }
|
| 227 |
+
if$
|
| 228 |
+
}
|
| 229 |
+
if$
|
| 230 |
+
}
|
| 231 |
+
't
|
| 232 |
+
if$
|
| 233 |
+
nameptr #1 + 'nameptr :=
|
| 234 |
+
|
| 235 |
+
namesleft #1 - 'namesleft :=
|
| 236 |
+
}
|
| 237 |
+
while$
|
| 238 |
+
}
|
| 239 |
+
|
| 240 |
+
FUNCTION {format.authors}
|
| 241 |
+
{ author empty$
|
| 242 |
+
{ "" }
|
| 243 |
+
{ author format.names }
|
| 244 |
+
if$
|
| 245 |
+
}
|
| 246 |
+
|
| 247 |
+
FUNCTION {format.editors}
|
| 248 |
+
{ editor empty$
|
| 249 |
+
{ "" }
|
| 250 |
+
{ editor format.names
|
| 251 |
+
editor num.names$ #1 >
|
| 252 |
+
{ ", editors" * }
|
| 253 |
+
{ ", editor" * }
|
| 254 |
+
if$
|
| 255 |
+
}
|
| 256 |
+
if$
|
| 257 |
+
}
|
| 258 |
+
|
| 259 |
+
FUNCTION {format.title}
|
| 260 |
+
{ title empty$
|
| 261 |
+
{ "" }
|
| 262 |
+
{ title "t" change.case$ }
|
| 263 |
+
if$
|
| 264 |
+
}
|
| 265 |
+
|
| 266 |
+
FUNCTION {n.dashify}
|
| 267 |
+
{ 't :=
|
| 268 |
+
""
|
| 269 |
+
{ t empty$ not }
|
| 270 |
+
{ t #1 #1 substring$ "-" =
|
| 271 |
+
{ t #1 #2 substring$ "--" = not
|
| 272 |
+
{ "--" *
|
| 273 |
+
t #2 global.max$ substring$ 't :=
|
| 274 |
+
}
|
| 275 |
+
{ { t #1 #1 substring$ "-" = }
|
| 276 |
+
{ "-" *
|
| 277 |
+
t #2 global.max$ substring$ 't :=
|
| 278 |
+
}
|
| 279 |
+
while$
|
| 280 |
+
}
|
| 281 |
+
if$
|
| 282 |
+
}
|
| 283 |
+
{ t #1 #1 substring$ *
|
| 284 |
+
t #2 global.max$ substring$ 't :=
|
| 285 |
+
}
|
| 286 |
+
if$
|
| 287 |
+
}
|
| 288 |
+
while$
|
| 289 |
+
}
|
| 290 |
+
|
| 291 |
+
FUNCTION {format.date}
|
| 292 |
+
{ year empty$
|
| 293 |
+
{ month empty$
|
| 294 |
+
{ "" }
|
| 295 |
+
{ "there's a month but no year in " cite$ * warning$
|
| 296 |
+
month
|
| 297 |
+
}
|
| 298 |
+
if$
|
| 299 |
+
}
|
| 300 |
+
{ month empty$
|
| 301 |
+
'year
|
| 302 |
+
{ month " " * year * }
|
| 303 |
+
if$
|
| 304 |
+
}
|
| 305 |
+
if$
|
| 306 |
+
}
|
| 307 |
+
|
| 308 |
+
FUNCTION {format.btitle}
|
| 309 |
+
{ title emphasize
|
| 310 |
+
}
|
| 311 |
+
|
| 312 |
+
FUNCTION {tie.or.space.connect}
|
| 313 |
+
{ duplicate$ text.length$ #3 <
|
| 314 |
+
{ "~" }
|
| 315 |
+
{ " " }
|
| 316 |
+
if$
|
| 317 |
+
swap$ * *
|
| 318 |
+
}
|
| 319 |
+
|
| 320 |
+
FUNCTION {either.or.check}
|
| 321 |
+
{ empty$
|
| 322 |
+
'pop$
|
| 323 |
+
{ "can't use both " swap$ * " fields in " * cite$ * warning$ }
|
| 324 |
+
if$
|
| 325 |
+
}
|
| 326 |
+
|
| 327 |
+
FUNCTION {format.bvolume}
|
| 328 |
+
{ volume empty$
|
| 329 |
+
{ "" }
|
| 330 |
+
{ "volume" volume tie.or.space.connect
|
| 331 |
+
series empty$
|
| 332 |
+
'skip$
|
| 333 |
+
{ " of " * series emphasize * }
|
| 334 |
+
if$
|
| 335 |
+
"volume and number" number either.or.check
|
| 336 |
+
}
|
| 337 |
+
if$
|
| 338 |
+
}
|
| 339 |
+
|
| 340 |
+
FUNCTION {format.number.series}
|
| 341 |
+
{ volume empty$
|
| 342 |
+
{ number empty$
|
| 343 |
+
{ series field.or.null }
|
| 344 |
+
{ output.state mid.sentence =
|
| 345 |
+
{ "number" }
|
| 346 |
+
{ "Number" }
|
| 347 |
+
if$
|
| 348 |
+
number tie.or.space.connect
|
| 349 |
+
series empty$
|
| 350 |
+
{ "there's a number but no series in " cite$ * warning$ }
|
| 351 |
+
{ " in " * series * }
|
| 352 |
+
if$
|
| 353 |
+
}
|
| 354 |
+
if$
|
| 355 |
+
}
|
| 356 |
+
{ "" }
|
| 357 |
+
if$
|
| 358 |
+
}
|
| 359 |
+
|
| 360 |
+
FUNCTION {format.edition}
|
| 361 |
+
{ edition empty$
|
| 362 |
+
{ "" }
|
| 363 |
+
{ output.state mid.sentence =
|
| 364 |
+
{ edition "l" change.case$ " edition" * }
|
| 365 |
+
{ edition "t" change.case$ " edition" * }
|
| 366 |
+
if$
|
| 367 |
+
}
|
| 368 |
+
if$
|
| 369 |
+
}
|
| 370 |
+
|
| 371 |
+
INTEGERS { multiresult }
|
| 372 |
+
|
| 373 |
+
FUNCTION {multi.page.check}
|
| 374 |
+
{ 't :=
|
| 375 |
+
#0 'multiresult :=
|
| 376 |
+
{ multiresult not
|
| 377 |
+
t empty$ not
|
| 378 |
+
and
|
| 379 |
+
}
|
| 380 |
+
{ t #1 #1 substring$
|
| 381 |
+
duplicate$ "-" =
|
| 382 |
+
swap$ duplicate$ "," =
|
| 383 |
+
swap$ "+" =
|
| 384 |
+
or or
|
| 385 |
+
{ #1 'multiresult := }
|
| 386 |
+
{ t #2 global.max$ substring$ 't := }
|
| 387 |
+
if$
|
| 388 |
+
}
|
| 389 |
+
while$
|
| 390 |
+
multiresult
|
| 391 |
+
}
|
| 392 |
+
|
| 393 |
+
FUNCTION {format.pages}
|
| 394 |
+
{ pages empty$
|
| 395 |
+
{ "" }
|
| 396 |
+
{ pages multi.page.check
|
| 397 |
+
{ "pages" pages n.dashify tie.or.space.connect }
|
| 398 |
+
{ "page" pages tie.or.space.connect }
|
| 399 |
+
if$
|
| 400 |
+
}
|
| 401 |
+
if$
|
| 402 |
+
}
|
| 403 |
+
|
| 404 |
+
FUNCTION {format.vol.num.pages}
|
| 405 |
+
{ volume field.or.null
|
| 406 |
+
number empty$
|
| 407 |
+
'skip$
|
| 408 |
+
{ "(" number * ")" * *
|
| 409 |
+
volume empty$
|
| 410 |
+
{ "there's a number but no volume in " cite$ * warning$ }
|
| 411 |
+
'skip$
|
| 412 |
+
if$
|
| 413 |
+
}
|
| 414 |
+
if$
|
| 415 |
+
pages empty$
|
| 416 |
+
'skip$
|
| 417 |
+
{ duplicate$ empty$
|
| 418 |
+
{ pop$ format.pages }
|
| 419 |
+
{ ":" * pages n.dashify * }
|
| 420 |
+
if$
|
| 421 |
+
}
|
| 422 |
+
if$
|
| 423 |
+
}
|
| 424 |
+
|
| 425 |
+
FUNCTION {format.chapter.pages}
|
| 426 |
+
{ chapter empty$
|
| 427 |
+
'format.pages
|
| 428 |
+
{ type empty$
|
| 429 |
+
{ "chapter" }
|
| 430 |
+
{ type "l" change.case$ }
|
| 431 |
+
if$
|
| 432 |
+
chapter tie.or.space.connect
|
| 433 |
+
pages empty$
|
| 434 |
+
'skip$
|
| 435 |
+
{ ", " * format.pages * }
|
| 436 |
+
if$
|
| 437 |
+
}
|
| 438 |
+
if$
|
| 439 |
+
}
|
| 440 |
+
|
| 441 |
+
FUNCTION {format.in.ed.booktitle}
|
| 442 |
+
{ booktitle empty$
|
| 443 |
+
{ "" }
|
| 444 |
+
{ editor empty$
|
| 445 |
+
{ "In " booktitle emphasize * }
|
| 446 |
+
{ "In " format.editors * ", " * booktitle emphasize * }
|
| 447 |
+
if$
|
| 448 |
+
}
|
| 449 |
+
if$
|
| 450 |
+
}
|
| 451 |
+
|
| 452 |
+
FUNCTION {empty.misc.check}
|
| 453 |
+
|
| 454 |
+
{ author empty$ title empty$ howpublished empty$
|
| 455 |
+
month empty$ year empty$ note empty$
|
| 456 |
+
and and and and and
|
| 457 |
+
key empty$ not and
|
| 458 |
+
{ "all relevant fields are empty in " cite$ * warning$ }
|
| 459 |
+
'skip$
|
| 460 |
+
if$
|
| 461 |
+
}
|
| 462 |
+
|
| 463 |
+
FUNCTION {format.thesis.type}
|
| 464 |
+
{ type empty$
|
| 465 |
+
'skip$
|
| 466 |
+
{ pop$
|
| 467 |
+
type "t" change.case$
|
| 468 |
+
}
|
| 469 |
+
if$
|
| 470 |
+
}
|
| 471 |
+
|
| 472 |
+
FUNCTION {format.tr.number}
|
| 473 |
+
{ type empty$
|
| 474 |
+
{ "Technical Report" }
|
| 475 |
+
'type
|
| 476 |
+
if$
|
| 477 |
+
number empty$
|
| 478 |
+
{ "t" change.case$ }
|
| 479 |
+
{ number tie.or.space.connect }
|
| 480 |
+
if$
|
| 481 |
+
}
|
| 482 |
+
|
| 483 |
+
FUNCTION {format.article.crossref}
|
| 484 |
+
{ key empty$
|
| 485 |
+
{ journal empty$
|
| 486 |
+
{ "need key or journal for " cite$ * " to crossref " * crossref *
|
| 487 |
+
warning$
|
| 488 |
+
""
|
| 489 |
+
}
|
| 490 |
+
{ "In {\em " journal * "\/}" * }
|
| 491 |
+
if$
|
| 492 |
+
}
|
| 493 |
+
{ "In " key * }
|
| 494 |
+
if$
|
| 495 |
+
" \cite{" * crossref * "}" *
|
| 496 |
+
}
|
| 497 |
+
|
| 498 |
+
FUNCTION {format.crossref.editor}
|
| 499 |
+
{ editor #1 "{vv~}{ll}" format.name$
|
| 500 |
+
editor num.names$ duplicate$
|
| 501 |
+
#2 >
|
| 502 |
+
{ pop$ " et~al." * }
|
| 503 |
+
{ #2 <
|
| 504 |
+
'skip$
|
| 505 |
+
{ editor #2 "{ff }{vv }{ll}{ jj}" format.name$ "others" =
|
| 506 |
+
{ " et~al." * }
|
| 507 |
+
{ " and " * editor #2 "{vv~}{ll}" format.name$ * }
|
| 508 |
+
if$
|
| 509 |
+
}
|
| 510 |
+
if$
|
| 511 |
+
}
|
| 512 |
+
if$
|
| 513 |
+
}
|
| 514 |
+
|
| 515 |
+
FUNCTION {format.book.crossref}
|
| 516 |
+
{ volume empty$
|
| 517 |
+
{ "empty volume in " cite$ * "'s crossref of " * crossref * warning$
|
| 518 |
+
"In "
|
| 519 |
+
}
|
| 520 |
+
{ "Volume" volume tie.or.space.connect
|
| 521 |
+
" of " *
|
| 522 |
+
}
|
| 523 |
+
if$
|
| 524 |
+
editor empty$
|
| 525 |
+
editor field.or.null author field.or.null =
|
| 526 |
+
or
|
| 527 |
+
{ key empty$
|
| 528 |
+
{ series empty$
|
| 529 |
+
{ "need editor, key, or series for " cite$ * " to crossref " *
|
| 530 |
+
crossref * warning$
|
| 531 |
+
"" *
|
| 532 |
+
}
|
| 533 |
+
{ "{\em " * series * "\/}" * }
|
| 534 |
+
if$
|
| 535 |
+
}
|
| 536 |
+
{ key * }
|
| 537 |
+
if$
|
| 538 |
+
}
|
| 539 |
+
{ format.crossref.editor * }
|
| 540 |
+
if$
|
| 541 |
+
" \cite{" * crossref * "}" *
|
| 542 |
+
}
|
| 543 |
+
|
| 544 |
+
FUNCTION {format.incoll.inproc.crossref}
|
| 545 |
+
{ editor empty$
|
| 546 |
+
editor field.or.null author field.or.null =
|
| 547 |
+
or
|
| 548 |
+
{ key empty$
|
| 549 |
+
{ booktitle empty$
|
| 550 |
+
{ "need editor, key, or booktitle for " cite$ * " to crossref " *
|
| 551 |
+
crossref * warning$
|
| 552 |
+
""
|
| 553 |
+
}
|
| 554 |
+
{ "In {\em " booktitle * "\/}" * }
|
| 555 |
+
if$
|
| 556 |
+
}
|
| 557 |
+
{ "In " key * }
|
| 558 |
+
if$
|
| 559 |
+
}
|
| 560 |
+
{ "In " format.crossref.editor * }
|
| 561 |
+
if$
|
| 562 |
+
" \cite{" * crossref * "}" *
|
| 563 |
+
}
|
| 564 |
+
|
| 565 |
+
FUNCTION {article}
|
| 566 |
+
{ output.bibitem
|
| 567 |
+
format.authors "author" output.check
|
| 568 |
+
new.block
|
| 569 |
+
format.title "title" output.check
|
| 570 |
+
new.block
|
| 571 |
+
crossref missing$
|
| 572 |
+
{ journal emphasize "journal" output.check
|
| 573 |
+
format.vol.num.pages output
|
| 574 |
+
format.date "year" output.check
|
| 575 |
+
}
|
| 576 |
+
{ format.article.crossref output.nonnull
|
| 577 |
+
format.pages output
|
| 578 |
+
}
|
| 579 |
+
if$
|
| 580 |
+
new.block
|
| 581 |
+
note output
|
| 582 |
+
fin.entry
|
| 583 |
+
}
|
| 584 |
+
|
| 585 |
+
FUNCTION {book}
|
| 586 |
+
{ output.bibitem
|
| 587 |
+
author empty$
|
| 588 |
+
{ format.editors "author and editor" output.check }
|
| 589 |
+
{ format.authors output.nonnull
|
| 590 |
+
crossref missing$
|
| 591 |
+
{ "author and editor" editor either.or.check }
|
| 592 |
+
'skip$
|
| 593 |
+
if$
|
| 594 |
+
}
|
| 595 |
+
if$
|
| 596 |
+
new.block
|
| 597 |
+
format.btitle "title" output.check
|
| 598 |
+
crossref missing$
|
| 599 |
+
{ format.bvolume output
|
| 600 |
+
new.block
|
| 601 |
+
format.number.series output
|
| 602 |
+
new.sentence
|
| 603 |
+
publisher "publisher" output.check
|
| 604 |
+
address output
|
| 605 |
+
}
|
| 606 |
+
{ new.block
|
| 607 |
+
format.book.crossref output.nonnull
|
| 608 |
+
}
|
| 609 |
+
if$
|
| 610 |
+
format.edition output
|
| 611 |
+
format.date "year" output.check
|
| 612 |
+
new.block
|
| 613 |
+
note output
|
| 614 |
+
fin.entry
|
| 615 |
+
}
|
| 616 |
+
|
| 617 |
+
FUNCTION {booklet}
|
| 618 |
+
{ output.bibitem
|
| 619 |
+
format.authors output
|
| 620 |
+
new.block
|
| 621 |
+
format.title "title" output.check
|
| 622 |
+
howpublished address new.block.checkb
|
| 623 |
+
howpublished output
|
| 624 |
+
address output
|
| 625 |
+
format.date output
|
| 626 |
+
new.block
|
| 627 |
+
note output
|
| 628 |
+
fin.entry
|
| 629 |
+
}
|
| 630 |
+
|
| 631 |
+
FUNCTION {inbook}
|
| 632 |
+
{ output.bibitem
|
| 633 |
+
author empty$
|
| 634 |
+
{ format.editors "author and editor" output.check }
|
| 635 |
+
{ format.authors output.nonnull
|
| 636 |
+
|
| 637 |
+
crossref missing$
|
| 638 |
+
{ "author and editor" editor either.or.check }
|
| 639 |
+
'skip$
|
| 640 |
+
if$
|
| 641 |
+
}
|
| 642 |
+
if$
|
| 643 |
+
new.block
|
| 644 |
+
format.btitle "title" output.check
|
| 645 |
+
crossref missing$
|
| 646 |
+
{ format.bvolume output
|
| 647 |
+
format.chapter.pages "chapter and pages" output.check
|
| 648 |
+
new.block
|
| 649 |
+
format.number.series output
|
| 650 |
+
new.sentence
|
| 651 |
+
publisher "publisher" output.check
|
| 652 |
+
address output
|
| 653 |
+
}
|
| 654 |
+
{ format.chapter.pages "chapter and pages" output.check
|
| 655 |
+
new.block
|
| 656 |
+
format.book.crossref output.nonnull
|
| 657 |
+
}
|
| 658 |
+
if$
|
| 659 |
+
format.edition output
|
| 660 |
+
format.date "year" output.check
|
| 661 |
+
new.block
|
| 662 |
+
note output
|
| 663 |
+
fin.entry
|
| 664 |
+
}
|
| 665 |
+
|
| 666 |
+
FUNCTION {incollection}
|
| 667 |
+
{ output.bibitem
|
| 668 |
+
format.authors "author" output.check
|
| 669 |
+
new.block
|
| 670 |
+
format.title "title" output.check
|
| 671 |
+
new.block
|
| 672 |
+
crossref missing$
|
| 673 |
+
{ format.in.ed.booktitle "booktitle" output.check
|
| 674 |
+
format.bvolume output
|
| 675 |
+
format.number.series output
|
| 676 |
+
format.chapter.pages output
|
| 677 |
+
new.sentence
|
| 678 |
+
publisher "publisher" output.check
|
| 679 |
+
address output
|
| 680 |
+
format.edition output
|
| 681 |
+
format.date "year" output.check
|
| 682 |
+
}
|
| 683 |
+
{ format.incoll.inproc.crossref output.nonnull
|
| 684 |
+
format.chapter.pages output
|
| 685 |
+
}
|
| 686 |
+
if$
|
| 687 |
+
new.block
|
| 688 |
+
note output
|
| 689 |
+
fin.entry
|
| 690 |
+
}
|
| 691 |
+
|
| 692 |
+
FUNCTION {inproceedings}
|
| 693 |
+
{ output.bibitem
|
| 694 |
+
format.authors "author" output.check
|
| 695 |
+
new.block
|
| 696 |
+
format.title "title" output.check
|
| 697 |
+
new.block
|
| 698 |
+
crossref missing$
|
| 699 |
+
{ format.in.ed.booktitle "booktitle" output.check
|
| 700 |
+
format.bvolume output
|
| 701 |
+
format.number.series output
|
| 702 |
+
format.pages output
|
| 703 |
+
address empty$
|
| 704 |
+
{ organization publisher new.sentence.checkb
|
| 705 |
+
organization output
|
| 706 |
+
publisher output
|
| 707 |
+
format.date "year" output.check
|
| 708 |
+
}
|
| 709 |
+
{ address output.nonnull
|
| 710 |
+
format.date "year" output.check
|
| 711 |
+
new.sentence
|
| 712 |
+
organization output
|
| 713 |
+
publisher output
|
| 714 |
+
}
|
| 715 |
+
if$
|
| 716 |
+
}
|
| 717 |
+
{ format.incoll.inproc.crossref output.nonnull
|
| 718 |
+
format.pages output
|
| 719 |
+
}
|
| 720 |
+
if$
|
| 721 |
+
new.block
|
| 722 |
+
note output
|
| 723 |
+
fin.entry
|
| 724 |
+
}
|
| 725 |
+
|
| 726 |
+
FUNCTION {conference} { inproceedings }
|
| 727 |
+
|
| 728 |
+
FUNCTION {manual}
|
| 729 |
+
{ output.bibitem
|
| 730 |
+
author empty$
|
| 731 |
+
{ organization empty$
|
| 732 |
+
'skip$
|
| 733 |
+
{ organization output.nonnull
|
| 734 |
+
address output
|
| 735 |
+
}
|
| 736 |
+
if$
|
| 737 |
+
}
|
| 738 |
+
{ format.authors output.nonnull }
|
| 739 |
+
if$
|
| 740 |
+
new.block
|
| 741 |
+
format.btitle "title" output.check
|
| 742 |
+
author empty$
|
| 743 |
+
{ organization empty$
|
| 744 |
+
{ address new.block.checka
|
| 745 |
+
address output
|
| 746 |
+
}
|
| 747 |
+
'skip$
|
| 748 |
+
if$
|
| 749 |
+
}
|
| 750 |
+
{ organization address new.block.checkb
|
| 751 |
+
organization output
|
| 752 |
+
address output
|
| 753 |
+
}
|
| 754 |
+
if$
|
| 755 |
+
format.edition output
|
| 756 |
+
format.date output
|
| 757 |
+
new.block
|
| 758 |
+
note output
|
| 759 |
+
fin.entry
|
| 760 |
+
}
|
| 761 |
+
|
| 762 |
+
FUNCTION {mastersthesis}
|
| 763 |
+
{ output.bibitem
|
| 764 |
+
format.authors "author" output.check
|
| 765 |
+
new.block
|
| 766 |
+
format.title "title" output.check
|
| 767 |
+
new.block
|
| 768 |
+
"Master's thesis" format.thesis.type output.nonnull
|
| 769 |
+
school "school" output.check
|
| 770 |
+
address output
|
| 771 |
+
format.date "year" output.check
|
| 772 |
+
new.block
|
| 773 |
+
note output
|
| 774 |
+
fin.entry
|
| 775 |
+
}
|
| 776 |
+
|
| 777 |
+
FUNCTION {misc}
|
| 778 |
+
{ output.bibitem
|
| 779 |
+
format.authors output
|
| 780 |
+
title howpublished new.block.checkb
|
| 781 |
+
format.title output
|
| 782 |
+
howpublished new.block.checka
|
| 783 |
+
howpublished output
|
| 784 |
+
format.date output
|
| 785 |
+
new.block
|
| 786 |
+
note output
|
| 787 |
+
fin.entry
|
| 788 |
+
empty.misc.check
|
| 789 |
+
}
|
| 790 |
+
|
| 791 |
+
FUNCTION {phdthesis}
|
| 792 |
+
{ output.bibitem
|
| 793 |
+
format.authors "author" output.check
|
| 794 |
+
new.block
|
| 795 |
+
format.btitle "title" output.check
|
| 796 |
+
new.block
|
| 797 |
+
"PhD thesis" format.thesis.type output.nonnull
|
| 798 |
+
school "school" output.check
|
| 799 |
+
address output
|
| 800 |
+
format.date "year" output.check
|
| 801 |
+
new.block
|
| 802 |
+
note output
|
| 803 |
+
fin.entry
|
| 804 |
+
}
|
| 805 |
+
|
| 806 |
+
FUNCTION {proceedings}
|
| 807 |
+
{ output.bibitem
|
| 808 |
+
editor empty$
|
| 809 |
+
{ organization output }
|
| 810 |
+
{ format.editors output.nonnull }
|
| 811 |
+
|
| 812 |
+
if$
|
| 813 |
+
new.block
|
| 814 |
+
format.btitle "title" output.check
|
| 815 |
+
format.bvolume output
|
| 816 |
+
format.number.series output
|
| 817 |
+
address empty$
|
| 818 |
+
{ editor empty$
|
| 819 |
+
{ publisher new.sentence.checka }
|
| 820 |
+
{ organization publisher new.sentence.checkb
|
| 821 |
+
organization output
|
| 822 |
+
}
|
| 823 |
+
if$
|
| 824 |
+
publisher output
|
| 825 |
+
format.date "year" output.check
|
| 826 |
+
}
|
| 827 |
+
{ address output.nonnull
|
| 828 |
+
format.date "year" output.check
|
| 829 |
+
new.sentence
|
| 830 |
+
editor empty$
|
| 831 |
+
'skip$
|
| 832 |
+
{ organization output }
|
| 833 |
+
if$
|
| 834 |
+
publisher output
|
| 835 |
+
}
|
| 836 |
+
if$
|
| 837 |
+
new.block
|
| 838 |
+
note output
|
| 839 |
+
fin.entry
|
| 840 |
+
}
|
| 841 |
+
|
| 842 |
+
FUNCTION {techreport}
|
| 843 |
+
{ output.bibitem
|
| 844 |
+
format.authors "author" output.check
|
| 845 |
+
new.block
|
| 846 |
+
format.title "title" output.check
|
| 847 |
+
new.block
|
| 848 |
+
format.tr.number output.nonnull
|
| 849 |
+
institution "institution" output.check
|
| 850 |
+
address output
|
| 851 |
+
format.date "year" output.check
|
| 852 |
+
new.block
|
| 853 |
+
note output
|
| 854 |
+
fin.entry
|
| 855 |
+
}
|
| 856 |
+
|
| 857 |
+
FUNCTION {unpublished}
|
| 858 |
+
{ output.bibitem
|
| 859 |
+
format.authors "author" output.check
|
| 860 |
+
new.block
|
| 861 |
+
format.title "title" output.check
|
| 862 |
+
new.block
|
| 863 |
+
note "note" output.check
|
| 864 |
+
format.date output
|
| 865 |
+
fin.entry
|
| 866 |
+
}
|
| 867 |
+
|
| 868 |
+
FUNCTION {default.type} { misc }
|
| 869 |
+
|
| 870 |
+
MACRO {jan} {"Jan."}
|
| 871 |
+
|
| 872 |
+
MACRO {feb} {"Feb."}
|
| 873 |
+
|
| 874 |
+
MACRO {mar} {"Mar."}
|
| 875 |
+
|
| 876 |
+
MACRO {apr} {"Apr."}
|
| 877 |
+
|
| 878 |
+
MACRO {may} {"May"}
|
| 879 |
+
|
| 880 |
+
MACRO {jun} {"June"}
|
| 881 |
+
|
| 882 |
+
MACRO {jul} {"July"}
|
| 883 |
+
|
| 884 |
+
MACRO {aug} {"Aug."}
|
| 885 |
+
|
| 886 |
+
MACRO {sep} {"Sept."}
|
| 887 |
+
|
| 888 |
+
MACRO {oct} {"Oct."}
|
| 889 |
+
|
| 890 |
+
MACRO {nov} {"Nov."}
|
| 891 |
+
|
| 892 |
+
MACRO {dec} {"Dec."}
|
| 893 |
+
|
| 894 |
+
MACRO {acmcs} {"ACM Comput. Surv."}
|
| 895 |
+
|
| 896 |
+
MACRO {acta} {"Acta Inf."}
|
| 897 |
+
|
| 898 |
+
MACRO {cacm} {"Commun. ACM"}
|
| 899 |
+
|
| 900 |
+
MACRO {ibmjrd} {"IBM J. Res. Dev."}
|
| 901 |
+
|
| 902 |
+
MACRO {ibmsj} {"IBM Syst.~J."}
|
| 903 |
+
|
| 904 |
+
MACRO {ieeese} {"IEEE Trans. Softw. Eng."}
|
| 905 |
+
|
| 906 |
+
MACRO {ieeetc} {"IEEE Trans. Comput."}
|
| 907 |
+
|
| 908 |
+
MACRO {ieeetcad}
|
| 909 |
+
{"IEEE Trans. Comput.-Aided Design Integrated Circuits"}
|
| 910 |
+
|
| 911 |
+
MACRO {ipl} {"Inf. Process. Lett."}
|
| 912 |
+
|
| 913 |
+
MACRO {jacm} {"J.~ACM"}
|
| 914 |
+
|
| 915 |
+
MACRO {jcss} {"J.~Comput. Syst. Sci."}
|
| 916 |
+
|
| 917 |
+
MACRO {scp} {"Sci. Comput. Programming"}
|
| 918 |
+
|
| 919 |
+
MACRO {sicomp} {"SIAM J. Comput."}
|
| 920 |
+
|
| 921 |
+
MACRO {tocs} {"ACM Trans. Comput. Syst."}
|
| 922 |
+
|
| 923 |
+
MACRO {tods} {"ACM Trans. Database Syst."}
|
| 924 |
+
|
| 925 |
+
MACRO {tog} {"ACM Trans. Gr."}
|
| 926 |
+
|
| 927 |
+
MACRO {toms} {"ACM Trans. Math. Softw."}
|
| 928 |
+
|
| 929 |
+
MACRO {toois} {"ACM Trans. Office Inf. Syst."}
|
| 930 |
+
|
| 931 |
+
MACRO {toplas} {"ACM Trans. Prog. Lang. Syst."}
|
| 932 |
+
|
| 933 |
+
MACRO {tcs} {"Theoretical Comput. Sci."}
|
| 934 |
+
|
| 935 |
+
READ
|
| 936 |
+
|
| 937 |
+
FUNCTION {sortify}
|
| 938 |
+
{ purify$
|
| 939 |
+
"l" change.case$
|
| 940 |
+
}
|
| 941 |
+
|
| 942 |
+
INTEGERS { len }
|
| 943 |
+
|
| 944 |
+
FUNCTION {chop.word}
|
| 945 |
+
{ 's :=
|
| 946 |
+
'len :=
|
| 947 |
+
s #1 len substring$ =
|
| 948 |
+
{ s len #1 + global.max$ substring$ }
|
| 949 |
+
's
|
| 950 |
+
if$
|
| 951 |
+
}
|
| 952 |
+
|
| 953 |
+
FUNCTION {sort.format.names}
|
| 954 |
+
{ 's :=
|
| 955 |
+
#1 'nameptr :=
|
| 956 |
+
""
|
| 957 |
+
s num.names$ 'numnames :=
|
| 958 |
+
numnames 'namesleft :=
|
| 959 |
+
{ namesleft #0 > }
|
| 960 |
+
{ nameptr #1 >
|
| 961 |
+
{ " " * }
|
| 962 |
+
'skip$
|
| 963 |
+
if$
|
| 964 |
+
s nameptr "{vv{ } }{ll{ }}{ f{ }}{ jj{ }}" format.name$ 't :=
|
| 965 |
+
nameptr numnames = t "others" = and
|
| 966 |
+
{ "et al" * }
|
| 967 |
+
{ t sortify * }
|
| 968 |
+
if$
|
| 969 |
+
nameptr #1 + 'nameptr :=
|
| 970 |
+
namesleft #1 - 'namesleft :=
|
| 971 |
+
}
|
| 972 |
+
while$
|
| 973 |
+
}
|
| 974 |
+
|
| 975 |
+
FUNCTION {sort.format.title}
|
| 976 |
+
{ 't :=
|
| 977 |
+
"A " #2
|
| 978 |
+
"An " #3
|
| 979 |
+
"The " #4 t chop.word
|
| 980 |
+
chop.word
|
| 981 |
+
chop.word
|
| 982 |
+
sortify
|
| 983 |
+
#1 global.max$ substring$
|
| 984 |
+
}
|
| 985 |
+
|
| 986 |
+
FUNCTION {author.sort}
|
| 987 |
+
{ author empty$
|
| 988 |
+
{ key empty$
|
| 989 |
+
{ "to sort, need author or key in " cite$ * warning$
|
| 990 |
+
""
|
| 991 |
+
}
|
| 992 |
+
{ key sortify }
|
| 993 |
+
if$
|
| 994 |
+
}
|
| 995 |
+
{ author sort.format.names }
|
| 996 |
+
if$
|
| 997 |
+
}
|
| 998 |
+
|
| 999 |
+
FUNCTION {author.editor.sort}
|
| 1000 |
+
{ author empty$
|
| 1001 |
+
{ editor empty$
|
| 1002 |
+
{ key empty$
|
| 1003 |
+
{ "to sort, need author, editor, or key in " cite$ * warning$
|
| 1004 |
+
""
|
| 1005 |
+
}
|
| 1006 |
+
{ key sortify }
|
| 1007 |
+
if$
|
| 1008 |
+
}
|
| 1009 |
+
{ editor sort.format.names }
|
| 1010 |
+
if$
|
| 1011 |
+
}
|
| 1012 |
+
{ author sort.format.names }
|
| 1013 |
+
if$
|
| 1014 |
+
}
|
| 1015 |
+
|
| 1016 |
+
FUNCTION {author.organization.sort}
|
| 1017 |
+
{ author empty$
|
| 1018 |
+
|
| 1019 |
+
{ organization empty$
|
| 1020 |
+
{ key empty$
|
| 1021 |
+
{ "to sort, need author, organization, or key in " cite$ * warning$
|
| 1022 |
+
""
|
| 1023 |
+
}
|
| 1024 |
+
{ key sortify }
|
| 1025 |
+
if$
|
| 1026 |
+
}
|
| 1027 |
+
{ "The " #4 organization chop.word sortify }
|
| 1028 |
+
if$
|
| 1029 |
+
}
|
| 1030 |
+
{ author sort.format.names }
|
| 1031 |
+
if$
|
| 1032 |
+
}
|
| 1033 |
+
|
| 1034 |
+
FUNCTION {editor.organization.sort}
|
| 1035 |
+
{ editor empty$
|
| 1036 |
+
{ organization empty$
|
| 1037 |
+
{ key empty$
|
| 1038 |
+
{ "to sort, need editor, organization, or key in " cite$ * warning$
|
| 1039 |
+
""
|
| 1040 |
+
}
|
| 1041 |
+
{ key sortify }
|
| 1042 |
+
if$
|
| 1043 |
+
}
|
| 1044 |
+
{ "The " #4 organization chop.word sortify }
|
| 1045 |
+
if$
|
| 1046 |
+
}
|
| 1047 |
+
{ editor sort.format.names }
|
| 1048 |
+
if$
|
| 1049 |
+
}
|
| 1050 |
+
|
| 1051 |
+
FUNCTION {presort}
|
| 1052 |
+
{ type$ "book" =
|
| 1053 |
+
type$ "inbook" =
|
| 1054 |
+
or
|
| 1055 |
+
'author.editor.sort
|
| 1056 |
+
{ type$ "proceedings" =
|
| 1057 |
+
'editor.organization.sort
|
| 1058 |
+
{ type$ "manual" =
|
| 1059 |
+
'author.organization.sort
|
| 1060 |
+
'author.sort
|
| 1061 |
+
if$
|
| 1062 |
+
}
|
| 1063 |
+
if$
|
| 1064 |
+
}
|
| 1065 |
+
if$
|
| 1066 |
+
" "
|
| 1067 |
+
*
|
| 1068 |
+
year field.or.null sortify
|
| 1069 |
+
*
|
| 1070 |
+
" "
|
| 1071 |
+
*
|
| 1072 |
+
title field.or.null
|
| 1073 |
+
sort.format.title
|
| 1074 |
+
*
|
| 1075 |
+
#1 entry.max$ substring$
|
| 1076 |
+
'sort.key$ :=
|
| 1077 |
+
}
|
| 1078 |
+
|
| 1079 |
+
ITERATE {presort}
|
| 1080 |
+
|
| 1081 |
+
SORT
|
| 1082 |
+
|
| 1083 |
+
STRINGS { longest.label }
|
| 1084 |
+
|
| 1085 |
+
INTEGERS { number.label longest.label.width }
|
| 1086 |
+
|
| 1087 |
+
FUNCTION {initialize.longest.label}
|
| 1088 |
+
{ "" 'longest.label :=
|
| 1089 |
+
#1 'number.label :=
|
| 1090 |
+
#0 'longest.label.width :=
|
| 1091 |
+
}
|
| 1092 |
+
|
| 1093 |
+
FUNCTION {longest.label.pass}
|
| 1094 |
+
{ number.label int.to.str$ 'label :=
|
| 1095 |
+
number.label #1 + 'number.label :=
|
| 1096 |
+
label width$ longest.label.width >
|
| 1097 |
+
{ label 'longest.label :=
|
| 1098 |
+
label width$ 'longest.label.width :=
|
| 1099 |
+
}
|
| 1100 |
+
'skip$
|
| 1101 |
+
if$
|
| 1102 |
+
}
|
| 1103 |
+
|
| 1104 |
+
EXECUTE {initialize.longest.label}
|
| 1105 |
+
|
| 1106 |
+
ITERATE {longest.label.pass}
|
| 1107 |
+
|
| 1108 |
+
FUNCTION {begin.bib}
|
| 1109 |
+
{ preamble$ empty$
|
| 1110 |
+
'skip$
|
| 1111 |
+
{ preamble$ write$ newline$ }
|
| 1112 |
+
if$
|
| 1113 |
+
"\begin{thebibliography}{" longest.label * "}" *
|
| 1114 |
+
"\itemsep=-1pt" * % Compact the entries a little.
|
| 1115 |
+
write$ newline$
|
| 1116 |
+
}
|
| 1117 |
+
|
| 1118 |
+
EXECUTE {begin.bib}
|
| 1119 |
+
|
| 1120 |
+
EXECUTE {init.state.consts}
|
| 1121 |
+
|
| 1122 |
+
ITERATE {call.type$}
|
| 1123 |
+
|
| 1124 |
+
FUNCTION {end.bib}
|
| 1125 |
+
{ newline$
|
| 1126 |
+
"\end{thebibliography}" write$ newline$
|
| 1127 |
+
}
|
| 1128 |
+
|
| 1129 |
+
EXECUTE {end.bib}
|
| 1130 |
+
|
| 1131 |
+
% end of file ieee.bst
|
| 1132 |
+
% ---------------------------------------------------------------
|
| 1133 |
+
|
| 1134 |
+
|
| 1135 |
+
|
paper-tex/ms.bbl
ADDED
|
@@ -0,0 +1,337 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
\begin{thebibliography}{10}\itemsep=-1pt
|
| 2 |
+
|
| 3 |
+
\bibitem{Pixel6as49}
|
| 4 |
+
Pixel 6, a smarter chip for a smarter phone - google store.
|
| 5 |
+
\newblock \url{https://store.google.com/product/pixel_6?hl=en-US}.
|
| 6 |
+
\newblock (Accessed on 11/14/2021).
|
| 7 |
+
|
| 8 |
+
\bibitem{ttpla}
|
| 9 |
+
Rabab Abdelfattah, Xiaofeng Wang, and Song Wang.
|
| 10 |
+
\newblock Ttpla: An aerial-image dataset for detection and segmentation of
|
| 11 |
+
transmission towers and power lines.
|
| 12 |
+
\newblock In {\em Proceedings of the Asian Conference on Computer Vision},
|
| 13 |
+
2020.
|
| 14 |
+
|
| 15 |
+
\bibitem{barnes2009patchmatch}
|
| 16 |
+
Connelly Barnes, Eli Shechtman, Adam Finkelstein, and Dan~B Goldman.
|
| 17 |
+
\newblock Patchmatch: A randomized correspondence algorithm for structural
|
| 18 |
+
image editing.
|
| 19 |
+
\newblock {\em ACM Trans. Graph.}, 28(3):24, 2009.
|
| 20 |
+
|
| 21 |
+
\bibitem{deeplab}
|
| 22 |
+
Liang-Chieh Chen, George Papandreou, Iasonas Kokkinos, Kevin Murphy, and Alan~L
|
| 23 |
+
Yuille.
|
| 24 |
+
\newblock Deeplab: Semantic image segmentation with deep convolutional nets,
|
| 25 |
+
atrous convolution, and fully connected crfs.
|
| 26 |
+
\newblock {\em IEEE transactions on pattern analysis and machine intelligence},
|
| 27 |
+
40(4):834--848, 2017.
|
| 28 |
+
|
| 29 |
+
\bibitem{deeplabv3}
|
| 30 |
+
Liang-Chieh Chen, George Papandreou, Florian Schroff, and Hartwig Adam.
|
| 31 |
+
\newblock Rethinking atrous convolution for semantic image segmentation.
|
| 32 |
+
\newblock {\em arXiv preprint arXiv:1706.05587}, 2017.
|
| 33 |
+
|
| 34 |
+
\bibitem{deeplabv3p}
|
| 35 |
+
Liang-Chieh Chen, Yukun Zhu, George Papandreou, Florian Schroff, and Hartwig
|
| 36 |
+
Adam.
|
| 37 |
+
\newblock Encoder-decoder with atrous separable convolution for semantic image
|
| 38 |
+
segmentation.
|
| 39 |
+
\newblock In {\em Proceedings of the European conference on computer vision
|
| 40 |
+
(ECCV)}, pages 801--818, 2018.
|
| 41 |
+
|
| 42 |
+
\bibitem{glnet}
|
| 43 |
+
Wuyang Chen, Ziyu Jiang, Zhangyang Wang, Kexin Cui, and Xiaoning Qian.
|
| 44 |
+
\newblock Collaborative global-local networks for memory-efficient segmentation
|
| 45 |
+
of ultra-high resolution images.
|
| 46 |
+
\newblock In {\em Proceedings of the IEEE/CVF Conference on Computer Vision and
|
| 47 |
+
Pattern Recognition}, pages 8924--8933, 2019.
|
| 48 |
+
|
| 49 |
+
\bibitem{cascadepsp}
|
| 50 |
+
Ho~Kei Cheng, Jihoon Chung, Yu-Wing Tai, and Chi-Keung Tang.
|
| 51 |
+
\newblock Cascadepsp: toward class-agnostic and very high-resolution
|
| 52 |
+
segmentation via global and local refinement.
|
| 53 |
+
\newblock In {\em Proceedings of the IEEE/CVF Conference on Computer Vision and
|
| 54 |
+
Pattern Recognition}, pages 8890--8899, 2020.
|
| 55 |
+
|
| 56 |
+
\bibitem{darabi2012image}
|
| 57 |
+
Soheil Darabi, Eli Shechtman, Connelly Barnes, Dan~B Goldman, and Pradeep Sen.
|
| 58 |
+
\newblock Image melding: Combining inconsistent images using patch-based
|
| 59 |
+
synthesis.
|
| 60 |
+
\newblock {\em ACM Transactions on graphics (TOG)}, 31(4):1--10, 2012.
|
| 61 |
+
|
| 62 |
+
\bibitem{isdnet}
|
| 63 |
+
Shaohua Guo, Liang Liu, Zhenye Gan, Yabiao Wang, Wuhao Zhang, Chengjie Wang,
|
| 64 |
+
Guannan Jiang, Wei Zhang, Ran Yi, Lizhuang Ma, et~al.
|
| 65 |
+
\newblock Isdnet: Integrating shallow and deep networks for efficient
|
| 66 |
+
ultra-high resolution segmentation.
|
| 67 |
+
\newblock In {\em Proceedings of the IEEE/CVF Conference on Computer Vision and
|
| 68 |
+
Pattern Recognition}, pages 4361--4370, 2022.
|
| 69 |
+
|
| 70 |
+
\bibitem{hassani2022dilated}
|
| 71 |
+
Ali Hassani and Humphrey Shi.
|
| 72 |
+
\newblock Dilated neighborhood attention transformer.
|
| 73 |
+
\newblock 2022.
|
| 74 |
+
|
| 75 |
+
\bibitem{hassani2022neighborhood}
|
| 76 |
+
Ali Hassani, Steven Walton, Jiachen Li, Shen Li, and Humphrey Shi.
|
| 77 |
+
\newblock Neighborhood attention transformer.
|
| 78 |
+
\newblock 2022.
|
| 79 |
+
|
| 80 |
+
\bibitem{resnet}
|
| 81 |
+
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun.
|
| 82 |
+
\newblock Deep residual learning for image recognition.
|
| 83 |
+
\newblock In {\em Proceedings of the IEEE conference on computer vision and
|
| 84 |
+
pattern recognition}, pages 770--778, 2016.
|
| 85 |
+
|
| 86 |
+
\bibitem{ccnet}
|
| 87 |
+
Zilong Huang, Xinggang Wang, Lichao Huang, Chang Huang, Yunchao Wei, and Wenyu
|
| 88 |
+
Liu.
|
| 89 |
+
\newblock Ccnet: Criss-cross attention for semantic segmentation.
|
| 90 |
+
\newblock In {\em Proceedings of the IEEE/CVF International Conference on
|
| 91 |
+
Computer Vision}, pages 603--612, 2019.
|
| 92 |
+
|
| 93 |
+
\bibitem{magnet}
|
| 94 |
+
Chuong Huynh, Anh~Tuan Tran, Khoa Luu, and Minh Hoai.
|
| 95 |
+
\newblock Progressive semantic segmentation.
|
| 96 |
+
\newblock In {\em Proceedings of the IEEE/CVF Conference on Computer Vision and
|
| 97 |
+
Pattern Recognition}, pages 16755--16764, 2021.
|
| 98 |
+
|
| 99 |
+
\bibitem{globallocal}
|
| 100 |
+
Satoshi Iizuka, Edgar Simo-Serra, and Hiroshi Ishikawa.
|
| 101 |
+
\newblock Globally and locally consistent image completion.
|
| 102 |
+
\newblock {\em ACM Transactions on Graphics (ToG)}, 36(4):1--14, 2017.
|
| 103 |
+
|
| 104 |
+
\bibitem{jain2022oneformer}
|
| 105 |
+
Jitesh Jain, Jiachen Li, MangTik Chiu, Ali Hassani, Nikita Orlov, and Humphrey
|
| 106 |
+
Shi.
|
| 107 |
+
\newblock {OneFormer: One Transformer to Rule Universal Image Segmentation}.
|
| 108 |
+
\newblock 2023.
|
| 109 |
+
|
| 110 |
+
\bibitem{jain2021semask}
|
| 111 |
+
Jitesh Jain, Anukriti Singh, Nikita Orlov, Zilong Huang, Jiachen Li, Steven
|
| 112 |
+
Walton, and Humphrey Shi.
|
| 113 |
+
\newblock Semask: Semantically masking transformer backbones for effective
|
| 114 |
+
semantic segmentation.
|
| 115 |
+
\newblock {\em arXiv}, 2021.
|
| 116 |
+
|
| 117 |
+
\bibitem{jain2022keys}
|
| 118 |
+
Jitesh Jain, Yuqian Zhou, Ning Yu, and Humphrey Shi.
|
| 119 |
+
\newblock Keys to better image inpainting: Structure and texture go hand in
|
| 120 |
+
hand.
|
| 121 |
+
\newblock {\em arXiv preprint arXiv:2208.03382}, 2022.
|
| 122 |
+
|
| 123 |
+
\bibitem{swiftlane}
|
| 124 |
+
Oshada Jayasinghe, Damith Anhettigama, Sahan Hemachandra, Shenali Kariyawasam,
|
| 125 |
+
Ranga Rodrigo, and Peshala Jayasekara.
|
| 126 |
+
\newblock Swiftlane: Towards fast and efficient lane detection.
|
| 127 |
+
\newblock {\em arXiv preprint arXiv:2110.11779}, 2021.
|
| 128 |
+
|
| 129 |
+
\bibitem{karras2020analyzing}
|
| 130 |
+
Tero Karras, Samuli Laine, Miika Aittala, Janne Hellsten, Jaakko Lehtinen, and
|
| 131 |
+
Timo Aila.
|
| 132 |
+
\newblock Analyzing and improving the image quality of stylegan.
|
| 133 |
+
\newblock In {\em Proceedings of the IEEE/CVF conference on computer vision and
|
| 134 |
+
pattern recognition}, pages 8110--8119, 2020.
|
| 135 |
+
|
| 136 |
+
\bibitem{kaspar2015self}
|
| 137 |
+
Alexandre Kaspar, Boris Neubert, Dani Lischinski, Mark Pauly, and Johannes
|
| 138 |
+
Kopf.
|
| 139 |
+
\newblock Self tuning texture optimization.
|
| 140 |
+
\newblock In {\em Computer Graphics Forum}, volume~34, pages 349--359. Wiley
|
| 141 |
+
Online Library, 2015.
|
| 142 |
+
|
| 143 |
+
\bibitem{cable_inst}
|
| 144 |
+
Bo Li, Cheng Chen, Shiwen Dong, and Junfeng Qiao.
|
| 145 |
+
\newblock Transmission line detection in aerial images: An instance
|
| 146 |
+
segmentation approach based on multitask neural networks.
|
| 147 |
+
\newblock {\em Signal Processing: Image Communication}, 96:116278, 2021.
|
| 148 |
+
|
| 149 |
+
\bibitem{focal}
|
| 150 |
+
Tsung-Yi Lin, Priya Goyal, Ross Girshick, Kaiming He, and Piotr Doll{\'a}r.
|
| 151 |
+
\newblock Focal loss for dense object detection.
|
| 152 |
+
\newblock In {\em Proceedings of the IEEE international conference on computer
|
| 153 |
+
vision}, pages 2980--2988, 2017.
|
| 154 |
+
|
| 155 |
+
\bibitem{partialconv}
|
| 156 |
+
Guilin Liu, Fitsum~A Reda, Kevin~J Shih, Ting-Chun Wang, Andrew Tao, and Bryan
|
| 157 |
+
Catanzaro.
|
| 158 |
+
\newblock Image inpainting for irregular holes using partial convolutions.
|
| 159 |
+
\newblock In {\em Proceedings of the European conference on computer vision
|
| 160 |
+
(ECCV)}, pages 85--100, 2018.
|
| 161 |
+
|
| 162 |
+
\bibitem{swin}
|
| 163 |
+
Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and
|
| 164 |
+
Baining Guo.
|
| 165 |
+
\newblock Swin transformer: Hierarchical vision transformer using shifted
|
| 166 |
+
windows.
|
| 167 |
+
\newblock {\em arXiv preprint arXiv:2103.14030}, 2021.
|
| 168 |
+
|
| 169 |
+
\bibitem{adamw}
|
| 170 |
+
Ilya Loshchilov and Frank Hutter.
|
| 171 |
+
\newblock Decoupled weight decay regularization.
|
| 172 |
+
\newblock {\em arXiv preprint arXiv:1711.05101}, 2017.
|
| 173 |
+
|
| 174 |
+
\bibitem{lsnet}
|
| 175 |
+
Van~Nhan Nguyen, Robert Jenssen, and Davide Roverso.
|
| 176 |
+
\newblock Ls-net: Fast single-shot line-segment detector.
|
| 177 |
+
\newblock {\em arXiv preprint arXiv:1912.09532}, 2019.
|
| 178 |
+
|
| 179 |
+
\bibitem{contextencoder}
|
| 180 |
+
Deepak Pathak, Philipp Krahenbuhl, Jeff Donahue, Trevor Darrell, and Alexei~A
|
| 181 |
+
Efros.
|
| 182 |
+
\newblock Context encoders: Feature learning by inpainting.
|
| 183 |
+
\newblock In {\em Proceedings of the IEEE conference on computer vision and
|
| 184 |
+
pattern recognition}, pages 2536--2544, 2016.
|
| 185 |
+
|
| 186 |
+
\bibitem{dalle}
|
| 187 |
+
Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen.
|
| 188 |
+
\newblock Hierarchical text-conditional image generation with clip latents.
|
| 189 |
+
\newblock {\em arXiv preprint arXiv:2204.06125}, 2022.
|
| 190 |
+
|
| 191 |
+
\bibitem{dpt}
|
| 192 |
+
Ren{\'e} Ranftl, Alexey Bochkovskiy, and Vladlen Koltun.
|
| 193 |
+
\newblock Vision transformers for dense prediction.
|
| 194 |
+
\newblock In {\em Proceedings of the IEEE/CVF International Conference on
|
| 195 |
+
Computer Vision}, pages 12179--12188, 2021.
|
| 196 |
+
|
| 197 |
+
\bibitem{rombach2022high}
|
| 198 |
+
Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Bj{\"o}rn
|
| 199 |
+
Ommer.
|
| 200 |
+
\newblock High-resolution image synthesis with latent diffusion models.
|
| 201 |
+
\newblock In {\em Proceedings of the IEEE/CVF Conference on Computer Vision and
|
| 202 |
+
Pattern Recognition}, pages 10684--10695, 2022.
|
| 203 |
+
|
| 204 |
+
\bibitem{ohem}
|
| 205 |
+
Abhinav Shrivastava, Abhinav Gupta, and Ross Girshick.
|
| 206 |
+
\newblock Training region-based object detectors with online hard example
|
| 207 |
+
mining.
|
| 208 |
+
\newblock In {\em Proceedings of the IEEE conference on computer vision and
|
| 209 |
+
pattern recognition}, pages 761--769, 2016.
|
| 210 |
+
|
| 211 |
+
\bibitem{structurelane}
|
| 212 |
+
Jinming Su, Chao Chen, Ke Zhang, Junfeng Luo, Xiaoming Wei, and Xiaolin Wei.
|
| 213 |
+
\newblock Structure guided lane detection.
|
| 214 |
+
\newblock {\em arXiv preprint arXiv:2105.05403}, 2021.
|
| 215 |
+
|
| 216 |
+
\bibitem{suvorov2022resolution}
|
| 217 |
+
Roman Suvorov, Elizaveta Logacheva, Anton Mashikhin, Anastasia Remizova,
|
| 218 |
+
Arsenii Ashukha, Aleksei Silvestrov, Naejin Kong, Harshith Goka, Kiwoong
|
| 219 |
+
Park, and Victor Lempitsky.
|
| 220 |
+
\newblock Resolution-robust large mask inpainting with fourier convolutions.
|
| 221 |
+
\newblock In {\em Proceedings of the IEEE/CVF Winter Conference on Applications
|
| 222 |
+
of Computer Vision}, pages 2149--2159, 2022.
|
| 223 |
+
|
| 224 |
+
\bibitem{lanedet}
|
| 225 |
+
Lucas Tabelini, Rodrigo Berriel, Thiago~M Paixao, Claudine Badue, Alberto~F
|
| 226 |
+
De~Souza, and Thiago Oliveira-Santos.
|
| 227 |
+
\newblock Keep your eyes on the lane: Real-time attention-guided lane
|
| 228 |
+
detection.
|
| 229 |
+
\newblock In {\em Proceedings of the IEEE/CVF Conference on Computer Vision and
|
| 230 |
+
Pattern Recognition}, pages 294--302, 2021.
|
| 231 |
+
|
| 232 |
+
\bibitem{attention}
|
| 233 |
+
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones,
|
| 234 |
+
Aidan~N Gomez, {\L}ukasz Kaiser, and Illia Polosukhin.
|
| 235 |
+
\newblock Attention is all you need.
|
| 236 |
+
\newblock In {\em Advances in neural information processing systems}, pages
|
| 237 |
+
5998--6008, 2017.
|
| 238 |
+
|
| 239 |
+
\bibitem{wexler2007space}
|
| 240 |
+
Yonatan Wexler, Eli Shechtman, and Michal Irani.
|
| 241 |
+
\newblock Space-time completion of video.
|
| 242 |
+
\newblock {\em IEEE Transactions on pattern analysis and machine intelligence},
|
| 243 |
+
29(3):463--476, 2007.
|
| 244 |
+
|
| 245 |
+
\bibitem{segformer}
|
| 246 |
+
Enze Xie, Wenhai Wang, Zhiding Yu, Anima Anandkumar, Jose~M Alvarez, and Ping
|
| 247 |
+
Luo.
|
| 248 |
+
\newblock Segformer: Simple and efficient design for semantic segmentation with
|
| 249 |
+
transformers.
|
| 250 |
+
\newblock {\em arXiv preprint arXiv:2105.15203}, 2021.
|
| 251 |
+
|
| 252 |
+
\bibitem{xu2022image}
|
| 253 |
+
Xingqian Xu, Shant Navasardyan, Vahram Tadevosyan, Andranik Sargsyan, Yadong
|
| 254 |
+
Mu, and Humphrey Shi.
|
| 255 |
+
\newblock Image completion with heterogeneously filtered spectral hints.
|
| 256 |
+
\newblock In {\em WACV}, 2023.
|
| 257 |
+
|
| 258 |
+
\bibitem{powerlinedataset}
|
| 259 |
+
{\"O}mer~Emre Yetgin, {\"O}mer~Nezih Gerek, and {\"O}mer Nezih.
|
| 260 |
+
\newblock Power image dataset (infrared-ir and visible light-vl).
|
| 261 |
+
\newblock {\em Mendeley Data}, 8, 2017.
|
| 262 |
+
|
| 263 |
+
\bibitem{hifill}
|
| 264 |
+
Zili Yi, Qiang Tang, Shekoofeh Azizi, Daesik Jang, and Zhan Xu.
|
| 265 |
+
\newblock Contextual residual aggregation for ultra high-resolution image
|
| 266 |
+
inpainting.
|
| 267 |
+
\newblock In {\em Proceedings of the IEEE/CVF Conference on Computer Vision and
|
| 268 |
+
Pattern Recognition}, pages 7508--7517, 2020.
|
| 269 |
+
|
| 270 |
+
\bibitem{contextual}
|
| 271 |
+
Jiahui Yu, Zhe Lin, Jimei Yang, Xiaohui Shen, Xin Lu, and Thomas~S Huang.
|
| 272 |
+
\newblock Generative image inpainting with contextual attention.
|
| 273 |
+
\newblock In {\em Proceedings of the IEEE conference on computer vision and
|
| 274 |
+
pattern recognition}, pages 5505--5514, 2018.
|
| 275 |
+
|
| 276 |
+
\bibitem{yu2019free}
|
| 277 |
+
Jiahui Yu, Zhe Lin, Jimei Yang, Xiaohui Shen, Xin Lu, and Thomas~S Huang.
|
| 278 |
+
\newblock Free-form image inpainting with gated convolution.
|
| 279 |
+
\newblock In {\em Proceedings of the IEEE/CVF international conference on
|
| 280 |
+
computer vision}, pages 4471--4480, 2019.
|
| 281 |
+
|
| 282 |
+
\bibitem{zeng2020high}
|
| 283 |
+
Yu Zeng, Zhe Lin, Jimei Yang, Jianming Zhang, Eli Shechtman, and Huchuan Lu.
|
| 284 |
+
\newblock High-resolution image inpainting with iterative confidence feedback
|
| 285 |
+
and guided upsampling.
|
| 286 |
+
\newblock In {\em European conference on computer vision}, pages 1--17.
|
| 287 |
+
Springer, 2020.
|
| 288 |
+
|
| 289 |
+
\bibitem{pldu}
|
| 290 |
+
Heng Zhang, Wen Yang, Huai Yu, Haijian Zhang, and Gui-Song Xia.
|
| 291 |
+
\newblock Detecting power lines in uav images with convolutional features and
|
| 292 |
+
structured constraints.
|
| 293 |
+
\newblock {\em Remote Sensing}, 11(11):1342, 2019.
|
| 294 |
+
|
| 295 |
+
\bibitem{supercaf}
|
| 296 |
+
Lingzhi Zhang, Connelly Barnes, Kevin Wampler, Sohrab Amirghodsi, Eli
|
| 297 |
+
Shechtman, Zhe Lin, and Jianbo Shi.
|
| 298 |
+
\newblock Inpainting at modern camera resolution by guided patchmatch with
|
| 299 |
+
auto-curation.
|
| 300 |
+
\newblock In {\em European Conference on Computer Vision}, pages 51--67.
|
| 301 |
+
Springer, 2022.
|
| 302 |
+
|
| 303 |
+
\bibitem{pspnet}
|
| 304 |
+
Hengshuang Zhao, Jianping Shi, Xiaojuan Qi, Xiaogang Wang, and Jiaya Jia.
|
| 305 |
+
\newblock Pyramid scene parsing network.
|
| 306 |
+
\newblock In {\em Proceedings of the IEEE conference on computer vision and
|
| 307 |
+
pattern recognition}, pages 2881--2890, 2017.
|
| 308 |
+
|
| 309 |
+
\bibitem{comodgan}
|
| 310 |
+
Shengyu Zhao, Jonathan Cui, Yilun Sheng, Yue Dong, Xiao Liang, Eric~I Chang,
|
| 311 |
+
and Yan Xu.
|
| 312 |
+
\newblock Large scale image completion via co-modulated generative adversarial
|
| 313 |
+
networks.
|
| 314 |
+
\newblock {\em arXiv preprint arXiv:2103.10428}, 2021.
|
| 315 |
+
|
| 316 |
+
\bibitem{zheng2022cm}
|
| 317 |
+
Haitian Zheng, Zhe Lin, Jingwan Lu, Scott Cohen, Eli Shechtman, Connelly
|
| 318 |
+
Barnes, Jianming Zhang, Ning Xu, Sohrab Amirghodsi, and Jiebo Luo.
|
| 319 |
+
\newblock Cm-gan: Image inpainting with cascaded modulation gan and
|
| 320 |
+
object-aware training.
|
| 321 |
+
\newblock {\em arXiv preprint arXiv:2203.11947}, 2022.
|
| 322 |
+
|
| 323 |
+
\bibitem{setr}
|
| 324 |
+
Sixiao Zheng, Jiachen Lu, Hengshuang Zhao, Xiatian Zhu, Zekun Luo, Yabiao Wang,
|
| 325 |
+
Yanwei Fu, Jianfeng Feng, Tao Xiang, Philip~HS Torr, et~al.
|
| 326 |
+
\newblock Rethinking semantic segmentation from a sequence-to-sequence
|
| 327 |
+
perspective with transformers.
|
| 328 |
+
\newblock In {\em Proceedings of the IEEE/CVF Conference on Computer Vision and
|
| 329 |
+
Pattern Recognition}, pages 6881--6890, 2021.
|
| 330 |
+
|
| 331 |
+
\bibitem{zhou2017places}
|
| 332 |
+
Bolei Zhou, Agata Lapedriza, Aditya Khosla, Aude Oliva, and Antonio Torralba.
|
| 333 |
+
\newblock Places: A 10 million image database for scene recognition.
|
| 334 |
+
\newblock {\em IEEE Transactions on Pattern Analysis and Machine Intelligence},
|
| 335 |
+
2017.
|
| 336 |
+
|
| 337 |
+
\end{thebibliography}
|
paper-tex/ms.tex
ADDED
|
@@ -0,0 +1,100 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
% CVPR 2023 Paper Template
|
| 2 |
+
% based on the CVPR template provided by Ming-Ming Cheng (https://github.com/MCG-NKU/CVPR_Template)
|
| 3 |
+
% modified and extended by Stefan Roth ([email protected])
|
| 4 |
+
|
| 5 |
+
\documentclass[10pt,twocolumn,letterpaper]{article}
|
| 6 |
+
|
| 7 |
+
%%%%%%%%% PAPER TYPE - PLEASE UPDATE FOR FINAL VERSION
|
| 8 |
+
% \usepackage[review]{cvpr} % To produce the REVIEW version
|
| 9 |
+
\usepackage{cvpr} % To produce the CAMERA-READY version
|
| 10 |
+
%\usepackage[pagenumbers]{cvpr} % To force page numbers, e.g. for an arXiv version
|
| 11 |
+
|
| 12 |
+
\usepackage[accsupp]{axessibility} % Improves PDF readability for those with disabilities.
|
| 13 |
+
|
| 14 |
+
% Include other packages here, before hyperref.
|
| 15 |
+
\usepackage[normalem]{ulem}
|
| 16 |
+
\usepackage{graphicx}
|
| 17 |
+
\usepackage{amsmath}
|
| 18 |
+
\usepackage{amssymb}
|
| 19 |
+
\usepackage{booktabs}
|
| 20 |
+
\usepackage{xcolor}
|
| 21 |
+
\usepackage{comment}
|
| 22 |
+
\usepackage{enumitem}
|
| 23 |
+
\newcommand{\todo}[1]{{\color{red}#1}}
|
| 24 |
+
|
| 25 |
+
\newcommand{\benchmark}{WireSegHR}
|
| 26 |
+
|
| 27 |
+
% It is strongly recommended to use hyperref, especially for the review version.
|
| 28 |
+
% hyperref with option pagebackref eases the reviewers' job.
|
| 29 |
+
% Please disable hyperref *only* if you encounter grave issues, e.g. with the
|
| 30 |
+
% file validation for the camera-ready version.
|
| 31 |
+
%
|
| 32 |
+
% If you comment hyperref and then uncomment it, you should delete
|
| 33 |
+
% ReviewTempalte.aux before re-running LaTeX.
|
| 34 |
+
% (Or just hit 'q' on the first LaTeX run, let it finish, and you
|
| 35 |
+
% should be clear).
|
| 36 |
+
\usepackage[pagebackref,breaklinks,colorlinks]{hyperref}
|
| 37 |
+
|
| 38 |
+
|
| 39 |
+
% Support for easy cross-referencing
|
| 40 |
+
\usepackage[capitalize]{cleveref}
|
| 41 |
+
\crefname{section}{Sec.}{Secs.}
|
| 42 |
+
\Crefname{section}{Section}{Sections}
|
| 43 |
+
\Crefname{table}{Table}{Tables}
|
| 44 |
+
\crefname{table}{Tab.}{Tabs.}
|
| 45 |
+
|
| 46 |
+
|
| 47 |
+
%%%%%%%%% PAPER ID - PLEASE UPDATE
|
| 48 |
+
\def\cvprPaperID{699} % *** Enter the CVPR Paper ID here
|
| 49 |
+
\def\confName{CVPR}
|
| 50 |
+
\def\confYear{2023}
|
| 51 |
+
|
| 52 |
+
\begin{document}
|
| 53 |
+
|
| 54 |
+
%%%%%%%%% TITLE - PLEASE UPDATE
|
| 55 |
+
% \title{Automatic Wire Segmentation for Inpainting}
|
| 56 |
+
% \title{Segmentation to the Extreme: \\A Large-Scale Wire Segmentation Dataset and a Pilot Study}
|
| 57 |
+
%\title{Segmentation to the Extreme: \\A Large-Scale High-Resolution Wire Segmentation Dataset and a Pilot Study}
|
| 58 |
+
%\title{WINE: WIre NEver Appeared in Your Photos}
|
| 59 |
+
\title{Automatic High Resolution Wire Segmentation and Removal}
|
| 60 |
+
|
| 61 |
+
\author{Mang Tik Chiu$^{1,2}$, Xuaner Zhang$^2$, Zijun Wei$^{2}$, Yuqian Zhou$^2$, Eli Shechtman$^2$, \\Connelly Barnes$^{2}$, Zhe Lin$^2$, Florian Kainz$^2$, Sohrab Amirghodsi$^2$, Humphrey Shi$^{1,3}$
|
| 62 |
+
\\
|
| 63 |
+
{\small $^1$UIUC, $^2$Adobe, $^3$University of Oregon}\\
|
| 64 |
+
{\small \textbf{\url{https://github.com/adobe-research/auto-wire-removal}}}}
|
| 65 |
+
|
| 66 |
+
\twocolumn[{%
|
| 67 |
+
\renewcommand\twocolumn[1][]{#1}%
|
| 68 |
+
\maketitle
|
| 69 |
+
\begin{center}
|
| 70 |
+
\centering
|
| 71 |
+
\captionsetup{type=figure}
|
| 72 |
+
\includegraphics[width=\textwidth]{figures/teaser2.pdf}
|
| 73 |
+
\captionof{figure}{We present an automatic high-resolution wire segmentation and removal pipeline. Each triad shows the high-resolution input image, our automatic wire segmentation result masked in red, and our full-resolution wire removal result. The visual quality of these photographs is greatly improved with our fully-automated wire clean-up system.}
|
| 74 |
+
\label{fig:demo}
|
| 75 |
+
\end{center}%
|
| 76 |
+
}]
|
| 77 |
+
|
| 78 |
+
% \vspace{3mm}
|
| 79 |
+
%[INTERNAL]: \href{https://session-0f3f6c51-53bd-446a-ad30-ea59ed8b6445.devbox.training.adobesensei.io/tensorboard/result.html}{Visualizations} of current models.
|
| 80 |
+
|
| 81 |
+
%%%%%%%%% ABSTRACT
|
| 82 |
+
% \input{figure_tex/teaser}
|
| 83 |
+
\input{sections/abstract}
|
| 84 |
+
|
| 85 |
+
%%%%%%%%% BODY TEXT
|
| 86 |
+
\input{sections/introduction}
|
| 87 |
+
\input{sections/related_work}
|
| 88 |
+
\input{sections/dataset}
|
| 89 |
+
\input{sections/method_yq}
|
| 90 |
+
\input{sections/results}
|
| 91 |
+
\input{sections/discussion}
|
| 92 |
+
\input{sections/conclusion}
|
| 93 |
+
|
| 94 |
+
%%%%%%%%% REFERENCES
|
| 95 |
+
{\small
|
| 96 |
+
\bibliographystyle{ieee_fullname}
|
| 97 |
+
\bibliography{egbib}
|
| 98 |
+
}
|
| 99 |
+
|
| 100 |
+
\end{document}
|
paper-tex/sections/abstract.tex
ADDED
|
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\begin{abstract}
|
| 3 |
+
\vspace{-5.5mm}
|
| 4 |
+
Wires and powerlines are common visual distractions that often undermine the aesthetics of photographs. The manual process of precisely segmenting and removing them is extremely tedious and may take up hours, especially on high-resolution photos where wires may span the entire space. In this paper, we present an automatic wire clean-up system that eases the process of wire segmentation and removal/inpainting to within a few seconds.
|
| 5 |
+
We observe several unique challenges: wires are thin, lengthy, and sparse. These are rare properties of subjects that common segmentation tasks cannot handle, especially in high-resolution images.
|
| 6 |
+
We thus propose a two-stage method that leverages both global and local contexts to accurately segment wires in high-resolution images efficiently, and a tile-based inpainting strategy to remove the wires given our predicted segmentation masks. We also introduce the first wire segmentation benchmark dataset, \benchmark. Finally, we demonstrate quantitatively and qualitatively that our wire clean-up system enables fully automated wire removal with great generalization to various wire appearances.
|
| 7 |
+
|
| 8 |
+
%We introduce the novel problem of semantic segmentation for wire-like objects in high-resolution images for photo retouching applications.
|
| 9 |
+
%We observe several challenging properties of wire appearances that are rare in most segmentation datasets -- thin, lengthy, and sparse, which recent works in high-resolution semantic segmentation cannot effectively solve. We thus propose a two-stage model that leverages global context and local information to predict accurate wire masks, and design an inference pipeline that efficiently handles high-resolution images.
|
| 10 |
+
%To encourage future research, we introduce WireSeg-HR, the first benchmark dataset on wire-like object semantic segmentation for photographic applications. Finally, we show our wire segmentation pipeline enables fully automated wire removal for photo retouching.
|
| 11 |
+
|
| 12 |
+
|
| 13 |
+
% , and show that our model outperforms baseline methods.
|
| 14 |
+
% In the end, we demonstrate that with our predicted wire masks, wire-like object removal, a common while tedious photo retouching step, can be made fully automated with high quality results.
|
| 15 |
+
|
| 16 |
+
% We introduce the novel problem of semantic segmentation for wire-like objects in high-resolution images for photo retouching applications.
|
| 17 |
+
% We observe several challenging properties of wire appearances that are rare in most segmentation datasets -- thin, extensive, sparse and discontinuous (due to frequent occlusions). We provide comprehensive analyses on the effect of common semantic segmentation approaches over these wire-like objects and propose a two-stage model to overcome these difficulties.
|
| 18 |
+
% To encourage research on tackling challenging properties of wires, we introduce WireSeg-HR, the first benchmark dataset on wire-like object semantic segmentation for photographic applications, and show that our model outperforms baseline methods.
|
| 19 |
+
% % In the end, we demonstrate that with our predicted wire masks, wire-like object removal, a common while tedious photo retouching step, can be made fully automated with high quality results.
|
| 20 |
+
|
| 21 |
+
|
| 22 |
+
\end{abstract}
|
| 23 |
+
% \vspace{-1.343cm}
|
paper-tex/sections/conclusion.tex
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\section{Conclusion}
|
| 3 |
+
|
| 4 |
+
In this paper, we propose a fully automated wire segmentation and removal system for high-resolution imagery. We demonstrate a segmentation method that maximally preserves sparse wire features and annotations, with a two-stage model that effectively uses global context and local details. The predicted segmentation mask is used in our tile-based wire inpainting model that has been demonstrated to produce seamless inpainting results.
|
| 5 |
+
We also introduce \benchmark, the first benchmark wire dataset with high-quality annotations. We hope our proposed method will provide insights into tackling semantic segmentation with high resolution image and annotation properties, and that our benchmark dataset encourage further research in wire segmentation and removal.
|
| 6 |
+
|
| 7 |
+
%\todo{ending}
|
| 8 |
+
% Our wire segmentation and removal system method is a key component in automatic wire removal, and encourages research in recognizing associated wire entities such as its shadows and reflections. Our work will also motivate research in efficient inpainting algorithms with a focus on challenging appearances of wires.
|
paper-tex/sections/dataset.tex
ADDED
|
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
|
| 3 |
+
\section{Dataset Collection and \benchmark} \label{sec:dataset}
|
| 4 |
+
% We present
|
| 5 |
+
%We will release the test set (407 images?) to provide a standardized platform for benchmarking. \yq{Should we mention here we only release the testing dataset, or just do not mention now whether we want to release the training one.}
|
| 6 |
+
%In this section, we describe the unique properties and provide statistics of WireSeg-HR, a high-resolution photographic image benchmark dataset for wire semantic segmentation.
|
| 7 |
+
|
| 8 |
+
\subsection{Image Source and Annotations}
|
| 9 |
+
|
| 10 |
+
% We collected high-resolution images with wires and wire-like objects from multiple sources. 20\% of the images are taken with DSLRs and smartphone cameras, while the others are sourced from the Internet (e.g. Flickr). As a result, the images have a variety of resolutions and have gone through different image processing pipelines. We target images with scenes that include city, street, rural area and landscape. We restricted the number of images in which the wire is the major object. We manually filter images so that wires in these images appear differently, for example, in front of buildings, partially occluded by vegetation or be hardly visible in the sky. This leads to a rich set of wire appearances. The final dataset after manual selection contains 3201 images.
|
| 11 |
+
Our definition of wires include electrical wires/cables, power lines, supporting/connecting wires, and any wire-like object that resemble a wire structure.
|
| 12 |
+
We collect high-resolution images with wires from two sources: 80\% of the images are from photo sharing platforms (Flickr, Pixabay, etc.), and 20\% of the images are captured with different cameras (DSLRs and smartphones) in multiple countries on our own. For the internet images, we collect 400K candidate images by keyword-searching. Then, we remove duplicates and images where wires are the only subjects. We then curate the final 6K images that cover sufficient scene diversity like city, street, rural area and landscape.
|
| 13 |
+
|
| 14 |
+
%\subsection{Wire annotation}
|
| 15 |
+
|
| 16 |
+
\input{figure_tex/annotation}
|
| 17 |
+
\input{figure_tex/pipeline}
|
| 18 |
+
Our wire annotation process contains two rounds. In the first round, annotators draw detailed masks over wires at full-resolution. The annotated masks enclose the main wire body and the boundary, oftentimes including a gradient falloff due to aliasing or defocus.
|
| 19 |
+
The boundary region annotation is crucial so as to avoid residual artifacts during wire removal.
|
| 20 |
+
In the second round, quality assurance is carried out to re-annotate unsatisfactory annotations. We show an example of our high-quality wire annotations in Figure~\ref{fig:annotation}.
|
| 21 |
+
% As shown, our dataset contains highly detailed masks for wires of various shapes and appearances.
|
| 22 |
+
|
| 23 |
+
\subsection{Dataset Statistics}
|
| 24 |
+
%\yq{Do we want to release the entire dataset. It's better to have a table comparing the statistics of different exisiting datasets and show the special cases we covered but others do not. }\cezhang{agreed, we should also mention and justify that this is the first wire dataset, unlike previous datasets such as the one for line detection or for satellite images, or any other available ones.}
|
| 25 |
+
|
| 26 |
+
In Table~\ref{table:stats}, we list the statistics of our dataset and compare them with existing wire-like datasets. Our dataset is the first wire dataset that contains high-resolution photographic images. The dataset is randomly split into 5000 training, 500 validation, and 500 testing images. We release 420 copyright-free test images with annotations.
|
| 27 |
+
%We will release the 500 test images (the \benchmark) with annotations upon paper acceptance.
|
| 28 |
+
|
| 29 |
+
%We will release \todo{how many?} test set as the first wire segmentation benchmark dataset.
|
| 30 |
+
|
| 31 |
+
\input{tables/stats}
|
| 32 |
+
|
paper-tex/sections/discussion.tex
ADDED
|
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
|
| 3 |
+
\section{Discussion}
|
| 4 |
+
|
| 5 |
+
% \subsection{Wire completion}
|
| 6 |
+
|
| 7 |
+
% In some cases, a part of a wire goes in front of a textured object and becomes difficult to recognize. When this happens, a light post-processing step can be performed to connect two wire segments and form a complete wire. We achieve this by fitting a parabola on each predicted wire segment. Pixels along the fitted curve that are originally predicted as background with a wire confidence of above 0.4 are reclassified as wire. We show in Figure~\ref{fig:completion} how this simple post-processing method can help picking up any missed wires due to blending with backgrounds.
|
| 8 |
+
|
| 9 |
+
\begin{comment}
|
| 10 |
+
\subsection{Wire removal}
|
| 11 |
+
%\input{figure_tex/removal}
|
| 12 |
+
|
| 13 |
+
Figure~\ref{fig:pixel6} (B) show two examples of automatic wire removal using wire segmentation masks from our proposed model. We use the ``Content-aware Fill''~\cite{barnes2009patchmatch} feature in Photoshop to remove wires. The full process of wire segmentation and removal is fully automatic without the need of any user interaction.
|
| 14 |
+
\end{comment}
|
| 15 |
+
|
| 16 |
+
% \subsection{Generalizing to large images}
|
| 17 |
+
|
| 18 |
+
% \input{figure_tex/panorama}
|
| 19 |
+
|
| 20 |
+
% Given the fact that wires are sparse in natural images, we show that our two-stage model generalizes to ultra-high resolution images such as panoramas without significant computation overhead. A $10$K by $2$K panorama with its segmentation result is shown in Figure~\ref{fig:panorama}. The full inference takes $6.7$ seconds and produces high-quality wire segmentation.
|
| 21 |
+
\vspace{-1.5mm}
|
| 22 |
+
\subsection{Comparison with Google Pixel 6}
|
| 23 |
+
\vspace{-1.5mm}
|
| 24 |
+
% \input{figure_tex/pixel6}
|
| 25 |
+
Recently, Google Pixel 6~\cite{Pixel6as49} announced the ``Magic Eraser'' photo feature that automatically detects and removes distractors. Note that this is a product feature and is not specifically designed for wires, and thus is hardly comparable with our method. We compare against this feature by uploading the images to Google Photos and applying ``Magic Eraser'' without manual intervention. We find that ``Magic Eraser'' performs well on wires with clear background, but it suffers from thin wires that are hardly visible and wires with complicated background. We show two examples in the supplementary material.
|
| 26 |
+
|
| 27 |
+
\subsection{Failure cases}
|
| 28 |
+
\vspace*{-2.5mm}
|
| 29 |
+
% \input{figure_tex/failure}
|
| 30 |
+
%\cezhang{todo: fill in caption for figure and revise this paragraph}
|
| 31 |
+
%Figure~\ref{fig:failure} shows examples where our model does not perform well. In the upper row example, the diagonal and vertical wires are ambiguously blended with the background texture, which make the wires hardly distinguishable. In the lower row case, thicker wires with complex structures rarely appear in the dataset, causing predictions of only part of the wire.
|
| 32 |
+
|
| 33 |
+
While our proposed wire segmentation model produces high-quality masks in most situations, there are still some challenging cases that our model cannot resolve. In particular, wires that are heavily blended in with surrounding structures/background, or wires under extreme lighting conditions are challenging to segment accurately. We show several examples in the supplementary material.
|
| 34 |
+
|
| 35 |
+
\input{tables/component}
|
| 36 |
+
\input{tables/thresholds}
|
| 37 |
+
\input{tables/inpaint}
|
| 38 |
+
\input{figure_tex/inpaint_compare}
|
| 39 |
+
|
| 40 |
+
% \subsection{Potential negative impacts}
|
| 41 |
+
|
| 42 |
+
% While our work is primarily used in image retouching applications, improper use of our wire segmentation model and content removal may spread misinformation for malicious intents. Content authentication could be used to mark the integrity of the shared content. Caution must be taken on the potential negative social impacts that come with the application of this work.
|
| 43 |
+
% \vspace{-2mm}
|
paper-tex/sections/introduction.tex
ADDED
|
@@ -0,0 +1,86 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
|
| 3 |
+
\section{Introduction}
|
| 4 |
+
|
| 5 |
+
% Distractor removal is a common step in photo editing. Conventionally, to remove any unwanted objects from an image, the user first draws a mask over the object, then uses a hole-filling algorithm to remove them and fill in the background. With recent advance in semantic segmentation, automatic masking of some common objects is possible. Among the types of distractors, wires are one of the most common yet unique objects to remove in photographs. Their ubiquity and unique attributes calls for extra attention and importance to perform automatic wire segmentation.
|
| 6 |
+
|
| 7 |
+
Oftentimes wire-like objects such as powerlines and cables can cross the full width of an image and ruin an otherwise beautiful composition. Removing these ``distractors'' is thus an essential step in photo retouching to improve the visual quality of a photograph. Conventionally, removing a wire-like object requires two steps: 1) segmenting out the wire-like object, and 2) removing the selected wire and inpainting with plausible contents. Both steps, if done manually, are extremely tedious and error-prone, especially for high-resolution photographs that may take photographers up to hours to reach a high-quality retouching result.
|
| 8 |
+
%The second step has benefited from existing image inpainting techniques (e.g. Content-aware Fill~\cite{barnes2009patchmatch} in Photoshop) and can be accomplished in an automatic way. The first step -- wire mask creation -- is much more manually intensive and error-prone, and automatic methods have yet to be fully explored.
|
| 9 |
+
|
| 10 |
+
% In this paper, we address the problem of wire masking as a way to automate the step of identifying and segmenting out those objects.
|
| 11 |
+
In this paper, we explore a fully-automated wire segmentation and inpainting solution for wire-like object segmentation and removal with tailored model architecture and data processing.
|
| 12 |
+
For simplicity, we use \textit{wire} to refer to all wire-like objects, including powerlines, cables, supporting/connecting wires, and objects with wire-like shapes.
|
| 13 |
+
|
| 14 |
+
Wire semantic segmentation has a seemingly similar problem setup with generic semantic segmentation tasks; they both take in a high-resolution image and generate dense predictions at a pixel level.
|
| 15 |
+
However, wire semantic segmentation bears a number of unique challenges. First, wires are commonly long and thin, oftentimes spanning the entire image yet having a diameter of only a handful of pixels. A few examples are shown in Figure~\ref{fig:motivation}. This prevents us from getting a precise mask based on regions of interest. Second, the input images can have arbitrarily high resolution up to 10k$\times$10k pixels for photographic retouching applications. Downsampling such high-resolution images can easily cause the thin wire structures to disappear. This poses a trade-off between preserving image size for inference quality and run-time efficiency.
|
| 16 |
+
Third, while wires have regular parabolic shapes, they are often partially occluded and can reappear at arbitrary image location, thus not continuous.
|
| 17 |
+
%This limits the use of parametric modeling
|
| 18 |
+
(e.g.~\cite{lanedet,swiftlane}).
|
| 19 |
+
%Segmentation predictions on downsampled images capture fewer details, whereas inference on high-resolution images requires quadratically more GPU memory that may be too big to fit into a modern GPU.
|
| 20 |
+
% We also note that recent semantic segmentation methods that specifically target high-resolution images also have their limitations when applied to wire images.
|
| 21 |
+
% We summarize the challenges of wires in Figure~\ref{fig:motivation}.
|
| 22 |
+
|
| 23 |
+
To account for these challenges, we propose a system for automatic wire semantic segmentation and removal. For segmentation, we design a two-stage coarse-to-fine model that leverages both pixel-level details in local patches and global semantics from the full image content, and runs efficiently at inference time. For inpainting, we adopt an efficient network architecture~\cite{suvorov2022resolution}, which enables us to use a tile-based approach to handle arbitrary high resolution. We design a training strategy to enforce color consistency between the inpainted region and the original image.
|
| 24 |
+
We also present the first benchmark dataset, \benchmark, for wire semantic segmentation tasks, where we collect and annotate high-resolution images with diverse scene contents and wire appearances.
|
| 25 |
+
We provide analyses and baseline comparisons to justify our design choices, which include data collection, augmentation, and our two-stage model design. Together, these design choices help us overcome the unique challenges of accurately segmenting wires.
|
| 26 |
+
% We report our model performance and efficiency on this benchmark dataset, and provide a set of comparisons with popular semantic segmentation methods on this task.
|
| 27 |
+
Our contributions are as follows:
|
| 28 |
+
% zwei{reordered the items}
|
| 29 |
+
|
| 30 |
+
\begin{itemize}[noitemsep]
|
| 31 |
+
%\item We introduce a novel task: wire semantic segmentation for high-resolution photographic images, an under-explored task that is extremely important for applications like photo retouching.
|
| 32 |
+
%\item We introduce a wire semantic segmentation benchmark dataset that consists of photographic images at high resolution, with large scene diversity, various types of cameras, a rich collection of wire shapes and high-definite manual annotations.
|
| 33 |
+
\item \textbf{Wire segmentation model:} We propose a two-stage model for wire semantic segmentation that leverages global context and local information to predict accurate wire masks at high resolution. We design an inference pipeline that can efficiently handle ultra-high resolution images.
|
| 34 |
+
\item \textbf{Wire inpainting strategy:} We design a tile-based inpainting strategy and tailor the inpainting method for our wire removal task given our segmentation results.
|
| 35 |
+
\item \textbf{\benchmark, a benchmark dataset:} We collect a wire segmentation benchmark dataset that consists of high resolution images, with diversity in wire shapes and scene contents. We also release the manual annotations that have been carefully curated to serve as ground truths. Besides, we also propose a benchmark dataset to evaluate inpainting quality.
|
| 36 |
+
%\yq{Explain the most important difference between ours and previous existing datasets like diversity of camera viewpoints or wire width etc. (mt:changed)}
|
| 37 |
+
%\yq{What is the advantage of the newly proposed model? Will it be helpful for processing ultra-high resolution images? Is the global information used efficiently? (mt:changed)}
|
| 38 |
+
%\item We describe a set of domain-specific data augmentation and inference pipelines specifically for wire-like objects, which help overcome unique challenges in this task and improve model performance.\yq{Do we have the ablation study to show the data augmentation improves the performance? If we want to claim it as contribution, it should have some benefits and novelty in the method itself. (mt: removed)}
|
| 39 |
+
\end{itemize}
|
| 40 |
+
|
| 41 |
+
%Distractor removal is a common component in photographic enhancement. Among the various distractors, wire is one of the most common and general types. Automatically detecting and segmenting wires in an image will greatly reduce the manual masking work. However,
|
| 42 |
+
% Conventionally, the user manually draws a mask over unwanted objects, then uses hole-filling algorithms to remove them. Recent advances in deep learning introduced object segmentation methods that automatically draw masks over objects of interest, thereby automating the masking process. Nevertheless,
|
| 43 |
+
% there are several unique challenges in training wire segmentation models: (a) the unique shape properties of wire (thin) makes it need special treatment such as resizing and sampling. This renders the use of common object segmentation methods in this task difficult, and require task-specific designs in model and training schemes. (b) it is hard to collect data (this is the reason why we have the synthetic data. (c) there is a lack of standard dataset to benchmark the progress.
|
| 44 |
+
|
| 45 |
+
%\textcolor{blue}{flow: distractor removal (common operation) -> manual masking -> automating masking (segmentation), among all distractors, wire unique and common and important and general -> automatkcill segment wire is important}
|
| 46 |
+
|
| 47 |
+
|
| 48 |
+
%\textcolor{blue}{wire seg bears similar to semantic segmentation in that (a, sinput, b..), but due to the attributes (thin, large span,,,) of wire, normal segmentation is suboptimal. to ward this end, we developped xxx that addresses these issues our models is a two stage corease file, maintains efficenty}.
|
| 49 |
+
|
| 50 |
+
% \textcolor{blue}{Moreover, we provide a systemactic frameworks from data data processing ,data ayucmetion to post-process}.
|
| 51 |
+
|
| 52 |
+
|
| 53 |
+
% \textcolor{blue}{realtive new field, wihtout too much data, previous benches marks (low...), we contrbute to this reserach field wtih a high quality benchmark, we also profiled a couple of metrics and discuss the performance }
|
| 54 |
+
|
| 55 |
+
%\textcolor{blue}{to sum up, our xxx 1, 2. 3}
|
| 56 |
+
|
| 57 |
+
|
| 58 |
+
|
| 59 |
+
|
| 60 |
+
% flow:
|
| 61 |
+
|
| 62 |
+
%\textcolor{red}{REARANGE SECTIONS TO ONE SMOOTH STORY}
|
| 63 |
+
|
| 64 |
+
%\subsection*{Conventional semantic segmentation}
|
| 65 |
+
%In the conventional setting of semantic segmentation, the model learns to segment entities such as person, table and car. A common trait of these objects of interest is that they are somewhat regularly shaped, which means the object's width and height are similar. This property is very advantageous in terms of model training, since the object features will remain equally prominent when the image is resized without distortion. As a result, it is a common approach to downsample a large image to fit within a GPU for training and inference without losing object details.
|
| 66 |
+
|
| 67 |
+
%\subsection*{Segmenting wires in outdoor images}
|
| 68 |
+
%Outdoor wires and cables exhibit significantly different shape and lighting properties than common objects. First, many wires in outdoor images are extremely long and thin. This means that wires can often span across the entire image, while being only a few pixels thick. This poses several problems to performing inference on images, since too much downsampling would cause the wire to disappear, while too little downsampling would impose significant computation overhead. Motivated by these challenges, we propose a pipeline for wire segmentation in high-resolution images for image enhancement and provide a set of baseline in this task. Our contributions are as follows:
|
| 69 |
+
|
| 70 |
+
%In this paper we systematically tackle the problem of wire segmentation and address issues including data collection, image pre-processing and data augmentation during training.
|
| 71 |
+
|
| 72 |
+
% \section{Introduction}
|
| 73 |
+
|
| 74 |
+
% Distractor removal is a common component in photographic enhancement. Conventionally, the user manually draws a mask over unwanted objects, then uses hole-filling algorithms to remove them. Recent advances in deep learning introduced object segmentation methods that automatically draw masks over objects of interest, thereby automating the masking process. Nevertheless, there are several unique challenges in wire segmentation that renders the use of common object segmentation methods in this task difficult, and require task-specific designs in model and training schemes.
|
| 75 |
+
|
| 76 |
+
% \subsection*{Conventional semantic segmentation}
|
| 77 |
+
% In the conventional setting of semantic segmentation, the model learns to segment entities such as person, table and car. A common trait of these objects of interest is that they are somewhat regularly shaped, which means the object's width and height are similar. This property is very advantageous in terms of model training, since the object features will remain equally prominent when the image is resized without distortion. As a result, it is a common approach to downsample a large image to fit within a GPU for training and inference without losing object details.
|
| 78 |
+
|
| 79 |
+
% \subsection*{Segmenting wires in outdoor images}
|
| 80 |
+
% Outdoor wires and cables exhibit significantly different shape and lighting properties than common objects. First, many wires in outdoor images are extremely long andthin. This means that wires can often span across the entire image, while being only a few pixels thick. This poses several problems to performing inference on images, since too much downsampling would cause the wire to disappear, while too little downsampling would impose significant computation overhead. Motivated by these challenges, we propose a pipeline for wire segmentation in high-resolution images for image enhancement and provide a set of baseline in this task. Our contributions are as follows:
|
| 81 |
+
|
| 82 |
+
% \begin{itemize}
|
| 83 |
+
% \item We introduce a wire semantic segmentation dataset consisting of high-resolution images with wires.
|
| 84 |
+
% \item We propose a two-stage model for wire segmentation that captures high-resolution image features without significant computation overhead.
|
| 85 |
+
% \item We describe a set of pre-processsing and post-processing pipelines specific for wires, which helps overcome unique challenges in this task.
|
| 86 |
+
% \end{itemize}
|
paper-tex/sections/method.tex
ADDED
|
@@ -0,0 +1,98 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
|
| 3 |
+
% \textcolor{blue}{ 2 options: 1. synthetic works: sec 3. data, sec 4. method, if synthetic not working: sec, method, sec, evaluation data collection: 1. coverage (indoor, outdoor, we even handshot...: multiple sources [flicker, google, etc, our own shooting]),
|
| 4 |
+
% multiple coverage (different keywords, select 400 from xxxK images, [ask mada to provde distritbuion, in supplemnetatyr]), 3. resolution: minimize. [How to convince pepople to use the test set..], annotation quality [in house annotation, multiple rounds of quality assurance]. We xx, 409 will be release}
|
| 5 |
+
|
| 6 |
+
% \textcolor{blue}{ if synthetic works, add another subsection in data section}
|
| 7 |
+
|
| 8 |
+
% \subsection{Generating synthetic dataset}
|
| 9 |
+
% While our dataset contains xxxx high-resolution images, the number is still insufficient when compared to other large segmentation datasets such as the well known Cityscapes~\cite{cityscapes}, Mapillary Vistas~\cite{mapillary} and MSCOCO dataset~\cite{coco}. To overcome the data insufficiency issue, one approach is to incorporate synthetic datasets with similar properties in the training. This is used in numerous vision related tasks such as 3D reconstruction~\cite{3drecon}, semantic segmentation (specifically domain adaptation)~\cite{domainadapt}. Synthetic dataset is also widely used in line and wire segmentation tasks~\cite{syntheticline, syntheticwire}, though they primarily used 3D rendering engines to generate artificial wires.
|
| 10 |
+
|
| 11 |
+
% Contrary to their approaches, we opt to generate our synthetic dataset with existing wires by blending wires from one image to another. Our approach is motivated by the fact that wires in our dataset have very different shapes and sizes, and is therefore difficult to generate artificially. To ensure our synthetic dataset is adequately realistic, we leverage semantic information of the scene as a cue to select regions for wire blending. For example, we only blend wires from the sky of one image to the sky of another image, such that the lighting, texture and nearby entities are similar.
|
| 12 |
+
|
| 13 |
+
% Here we describe in detail how we synthesize wire images. First, we predict a semantic map for each image in our real dataset. This gives us semantic information on the image and the wires. Then, we extract the wires from these images by cropping wire pixels and removing them from the background with Content-Aware Fill. For each obtained background image, we generate three separate images by randomly transforming (e.g. rotating and scaling) and blending several wires with the image, where the blending region is constricted by the type of object in the region and the selected wire. The resulting synthetic wire dataset contains xxxx images with different realistic wire patterns. Figure~\ref{fig:synthetic} shows an example of our generated high-resolution wire image.
|
| 14 |
+
|
| 15 |
+
|
| 16 |
+
\section{Method} \label{sec:method}
|
| 17 |
+
% Sec 4.1 Model structure
|
| 18 |
+
% Sec 4.2 how we combine: inference: end to end , information flow, how from an image to wire segmentation masks, inference method: 1. global run -> subset regions (detailed selection will be descried in 4.4.1) [describe the whole process -> (new para) explain how selected.]
|
| 19 |
+
% Sec 4.3 how we train: how we train to make usre share weights: [describe whole training, then detail each components ((new parap) pre-processing, batch training, sampling, etc)]
|
| 20 |
+
% Sec.4.4 pre-processing and post-processing
|
| 21 |
+
|
| 22 |
+
% In this section we[], the structure is in fig X
|
| 23 |
+
% 4.1: subsubection{global branch}: input, output, structurte, subsubsection{local branch}: input, output , structrure
|
| 24 |
+
% 4.2: combining branches, connecting, how its done to end-to-end
|
| 25 |
+
|
| 26 |
+
% \textcolor{blue}{our model is composed of of xxx parts: a global (Secxxx), a local. Add a figure}
|
| 27 |
+
|
| 28 |
+
\input{figure_tex/pipeline}
|
| 29 |
+
|
| 30 |
+
In this section, we describe our two-stage model architecture. Figure~\ref{fig:pipeline} shows the overall structure of our model. We will first discuss motivations of our two-stage design, then we describe our two-stage model architecture and training scheme in detail.
|
| 31 |
+
|
| 32 |
+
\subsection{Motivation}
|
| 33 |
+
|
| 34 |
+
As shown in Figure~\ref{fig:motivation}, we divide our model into two major components -- a coarse module and a fine module. Our two-stage model design is motivated by two observations. To start, wires in our dataset are extremely long and thin. As mentioned above, many wires in our images can span over several thousand pixels long while only several pixels across. Limited by the memory size of GPUs, we cannot simply pass the entire image spanning thousands of pixels at full-resolution to a model for inference. As a result, two separate modules are required, where the coarse module captures the entire wire and its surrounding contextual information at a lower resolution, and the fine module captures detailed textures of the wire at the original resolution.
|
| 35 |
+
|
| 36 |
+
Next, we observe that wire pixels are also very sparse, where a typical high-resolution image contains very small percentages of wire pixels. This means that we can also use the coarse module to guide the fine module on what regions to capture full-resolution details. This way, the fine module can predict segmentation masks only where there are wires, and therefore reduce computation time.
|
| 37 |
+
|
| 38 |
+
\subsection{The two-stage coarse to fine model}
|
| 39 |
+
|
| 40 |
+
Given a high-resolution image $I$, the coarse module of our two-stage model aims to recognize semantics of the entire image and pick up wire regions at a coarse level. Looking at the entire image allows the coarse module to capture more contextual information. To fit the large images into a GPU for training and inference, images are first bilinearly downsampled, then fed into the coarse module to predict a global logit map $Z_{glo}$. We use $Z_{glo}$ as a conditional input to guide the fine module in the next stage.
|
| 41 |
+
|
| 42 |
+
The fine module is conditioned on the output from the coarse module. This module takes in a local image patch $I_{loc}$ of the whole image at full resolution, the global logit map $Z_{glo}$ from the coarse module, and a binary location mask $M$ that sets the patch relative to the full image to $1$ and other regions to $0$. The fine module then predicts the local wire logit map $Z_{loc}$. Empirically, we find that concatenating the entire global logit map rather than cropping the logit map at the location of the image patch yields slightly improved results.
|
| 43 |
+
|
| 44 |
+
The designs of the coarse and fine modules are conceptually the same as those in GLNet~\cite{glnet} and MagNet~\cite{magnet}, where a global network is trained on entire downsampled images and a local network is trained on higher-resolution image patches conditioned on some global features. However, unlike GLNet, where intermediate features are shared between the global and local branch bidirectionally, we opt for a simpler late fusion by concatenating the logit map directly to the fine module. We also only use two stages instead of up to four stage as done in MagNet, since only a fine module at the highest resolution is adequate at refining annotations that are only several pixels thick, and additional intermediate stages can drastically increase inference time.
|
| 45 |
+
|
| 46 |
+
Our model can be trained end-to-end. To do so, in the training stage, we first randomly scale, rotate, horizontal flip and apply photometric distortion to the image. The global image $I_{glo}$ is generated by simply downsampling this augmented image. We then generate the local image patch $I_{loc}$ by randomly cropping a 512$\times$512 window from the augmented image that contains at least 1\% wire pixels. This helps balance between wire and background pixels. We find that this simple constrained cropping approach yields better performances than other well known balancing methods, including Focal loss~\cite{focal} and Online Hard Example Mining~\cite{ohem}.
|
| 47 |
+
|
| 48 |
+
We share the same feature extractor network between both the coarse and fine module, but train separate feature decoders. To account for the additional inputs to the fine module, we expand the input channels of the feature extractor from 3 to 5. The last two channels are set to 0 when passing an image through the coarse module, and set to the global logit map $Z_{glo}$ and the binary location map $M$ for the fine module. Both $Z_{glo}$ and $Z_{loc}$ are trained by computing Cross Entropy losses of the logit maps after softmax against their respective ground truth wire annotations. The final loss $L$ of our model is the sum of the global cross entropy $L_{glo}$ loss and local cross entropy loss $L_{loc}$. They are defined as follows:
|
| 49 |
+
|
| 50 |
+
\begin{equation}
|
| 51 |
+
\begin{aligned}
|
| 52 |
+
L_{glo} &= -\sum_{c\in C}log\ Softmax(Z_{glo})_c \\
|
| 53 |
+
L_{loc} &= -\sum_{c\in C}log\ Softmax(Z_{loc})_c \\
|
| 54 |
+
L &= L_{glo} + L_{loc}
|
| 55 |
+
\end{aligned}
|
| 56 |
+
\end{equation}
|
| 57 |
+
where $C$ is the number of categories. $C=2$ (wire, background) in our task.\\
|
| 58 |
+
|
| 59 |
+
To perform inference with our model, for each high-resolution image, we first feed the downsampled image to the coarse module, which is the same as the training step. We then compute the global wire segmentation map by taking the $\mathrm{argmax}$ over the two classes for each pixel. Local refinement is done by running a sliding window over the entire image, where the fine step is performed if there are more than some percentage $\alpha$ of wire pixels within the image window. By conditioning on the global wire probability, we save computation time in regions where there are no wires, and maintain segmentation quality at high-resolution when potential wires are discovered. By passing the global logit to the fine module, we also allow the coarse module to provide information to the fine module for better inference quality.
|
| 60 |
+
|
| 61 |
+
\subsection{Implementation details}
|
| 62 |
+
|
| 63 |
+
We use MixTransformer~\cite{segformer} as our shared feature extractor. We expand the backbone's input RGB channel to five channels to accept the logit map and binary location map during the local step. We define two separate MLP feature decoders from~\cite{segformer} for the coarse and fine modules respectively.
|
| 64 |
+
|
| 65 |
+
We train our model on our training set with 2796 high-resolution images. The model is trained for 40k iterations with a batch size of 8. Global images are downsampled to 512$\times$512. We feed the global images to the coarse module and obtain a two channel class logit map. A single 1$\times$1 convolution is used to transform this logit map into a single channel map, which is then concatenated with the local RGB image and binary location mask. The five-channel input is finally fed into the fine module to obtain the local logit map. We use AdamW~\cite{adamw} with a learning rate of $0.00006$ and weight decay of $0.01$. We use the "poly" LR scheduling rule with a factor of $1.0$.
|
| 66 |
+
|
| 67 |
+
In the testing stage, we set both the global image and local image patch size to 1024$\times$1024. Unless otherwise specified, we set the percentage for local refinement $\alpha$ to $0.01\%$.
|
| 68 |
+
|
| 69 |
+
We also train common semantic segmentation models on our dataset for comparison. For whole image models, we train the models on downsampled whole-images, as are done in our coarse module and most semantic segmentation methods. For sliding-window models, we train them on full-resolution image patches, and perform inference in a sliding window approach. All experiments are trained on 4 Nvidia V100 GPUs and tested on a single V100 GPU.
|
| 70 |
+
|
| 71 |
+
% A straight-forward approach towards wire segmentation is to train a common object segmentation model on our high-resolution wire images. To avoid memory issues with large images, we adapt conventional semantic segmentation training pipelines by randomly downsampling the images to between 1024 and 2048 pixels on the longer side, while maintaining aspect ratio. Random rotation is added to increase variety of wire shapes. Random horizontal flip, photometric distortion and channel normalization are then added. To obtain the final segmentation prediction, the predicted probability map is bilinearly upsampled to the original image size, where evaluation takes place. We name this method coarse semantic segmentation, since the image downsampling and prediction upsampling steps cause the final output to have lower resolution than the original image.
|
| 72 |
+
|
| 73 |
+
% An intuitive alternative to coarse segmentation is to train the model on full-resolution image patches, then perform inference in a sliding-window manner. This way, the model captures the most details and thus gives the tightest prediction. The trade-off with the sliding window method is the inference time. Since each full-resolution image has to be cropped into multiple patches and be predicted individually, multiple iterations of inference have to be performed, resulting in significantly prolonged inference time when images are several times larger. As a comparison, we train and evaluate the same model architecture as coarse semantic segmentation, but instead of downsampling images to below 2048 pixels per side, we simply crop the full-resolution images without downsampling. During inference, a full-resolution large image is cropped into multiple image patches of the same size, and each patch is predicted individually. The finaly prediction map is constructed by stitching the prediction patches together. We name this method fine semantic segmentation, since the prediction is not upsampled and thus gives the tightest segmentation map.
|
| 74 |
+
|
| 75 |
+
% We show in Section~\ref{sec:results} that this method gives significantly better results than coarse semantic segmentation, but requires much longer to run.
|
| 76 |
+
|
| 77 |
+
% \subsection{The coarse-to-fine pipeline}
|
| 78 |
+
% Ideally, to perform efficient yet accurate semantic segmentation on large-resolution images, a fusion of the above two methods is most desired. Intuitively, we can use the coarse segmentation map as a condition to determine the regions in the image that require fine segmentation. To achieve this, we can divide the both the image and the coarse segmentation map into patches, and perform fine segmentation based on the result of the coarse segmentation patch, such as the percentage of wire pixels in the patch.
|
| 79 |
+
|
| 80 |
+
% Intuitively, we can simply use the trained models from the above two methods as the "combined model" for coarse-to-fine inference. Yet, this design requires two separate models, and do not share information between them. An improvement over this approach is to train a single model with both downsampled images and full-resolution image patches, then run coarse-to-fine inference using the same model, but this still does not provide a way to share information between both models.
|
| 81 |
+
|
| 82 |
+
% To ensure efficient inference without duplicated model structures, as well as enable information passing from the coarse inference stage to the fine inference stage, propose a sequential global-to-local network that shares the feature extractor backbone but contains two separate decoder heads. The concept can be thought of as the opposite of~\cite{glnet}, where they have separate feature extractors for global and local images. As a result, our model can be trained in an end-to-end manner, while their model requires a three-stage training scheme. Our proposed method is similar to~\cite{cascadepsp} in terms of the sequential inference pipeline, but differs in two major aspects. First, ~\cite{cascadepsp} shares the entire network for global and local inference. Second, their work is designed for segmentation refinement, meaning that their model requires a separate segmentation prediction as input, resulting in two iterations for global inference. On the other hand, our model only requires a single step in global inference.
|
| 83 |
+
|
| 84 |
+
% \subsection{Model description}
|
| 85 |
+
|
| 86 |
+
% Here we describe our two-stage model design and training details. Our model consists of a single shared feature extractor and two separate decoders. In our experiments, we use Mix Transformer (MiT)~\cite{segformer} as the feature extractor, and the multi-layer perceptron (MLP) decoder proposed by the same authors as the decoder structure. As mentioned above, we duplicate the decoder such that one is used for global inference and the other is used for local inference.
|
| 87 |
+
|
| 88 |
+
% To enable information sharing between the global step and the local step, we extend the model input from three channels to five. The first three channels are the original RGB channels, and the last two channels are for the local step, and is therefore deactivated (i.e. all zeros) during the global step. In the local step, the fourth channels is the probability map of the entire downsampled image. The fifth channel is a binary map, where the location of the local image patch projected back onto the whole image is set as one, and the remaining regions as zero. Compared to~\cite{cascadepsp}, where only the probability map of the local image patch is appended to the input, using the probability map of the entire image enables the network to gain contextual information of the image patch. At the same time, the binary map allows the network to condition on the local probability, thus performing accurate segmentation refinement. An illustration of our model design can be seen in Figure.~\ref{fig:model}.
|
| 89 |
+
|
| 90 |
+
% \textbf{stress on the nice part of the model: shared weights, size, etc...}
|
| 91 |
+
|
| 92 |
+
|
| 93 |
+
% \section{Toward Collecting High-resolution Wire Dataset}
|
| 94 |
+
|
| 95 |
+
% \subsection{High-resolution Wire Dataset}
|
| 96 |
+
% Albeit having major incentives towards distractor removal, few works exist that tackle wire segmentation in high-resolution outdoor images. To this end, we present a dataset consisting of xxx very high-resolution outdoor images with wires and cables. Image sizes in our dataset varies from X$\times$Y to W$\times$Z. Wire images in our dataset not only include power lines and cables as in~\cite{ttpla}, but also electrical wires attached to buildings and other structures. This adds significant variety to the types of wires in terms of color and lighting. In addition, while ~\cite{ttpla} contains power line images mostly from close distances, wires in our dataset images can appear from very far distances, which makes them extremely difficult to recognize even with careful visual inspection, to extreme close-ups where textures are clearly visible. This extreme pattern variety makes it very difficult for the model to infer from a single visual cue, such as color gradients.
|
| 97 |
+
|
| 98 |
+
% Table.\ref{table:dataset} shows statistics of our dataset. As can be seen, our dataset contains images that are significantly larger than other similar wire/cable datasets. (add stuff when stats and comparisons are done)
|
paper-tex/sections/method_yq.tex
ADDED
|
@@ -0,0 +1,111 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
|
| 3 |
+
% \section{Method} \label{sec:method}
|
| 4 |
+
|
| 5 |
+
% In this section, we describe our two-stage model architecture. Figure~\ref{fig:pipeline} shows the overall model structure. We will first discuss motivations of our two-stage design, then describe our model architecture and training details.
|
| 6 |
+
|
| 7 |
+
%In this section, we first discuss motivations of our two-stage model architecture, then describe our model architecture and training details. Figure~\ref{fig:pipeline} shows the overall model structure.
|
| 8 |
+
|
| 9 |
+
\section{High-Resolution Wire Segmentation}
|
| 10 |
+
\label{sec:segmentation}
|
| 11 |
+
Wires appear visually different from common objects -- being thin, long, sparse and oftentimes partially occluded.
|
| 12 |
+
We find the following two design choices crucial to building an effective wire segmentation system: 1) having a two stage framework so that coarse prediction from global context guides precise segmentation from local patches and 2) maximally preserving and augmenting image features and annotations of wires throughout the pipeline.
|
| 13 |
+
% We thus propose a two-stage coarse-to-fine pipeline, in which we inject feature augmentation and architectural choices to maximally preserve the wire features throughout.
|
| 14 |
+
|
| 15 |
+
\subsection{The Two-stage Coarse to Fine Model}
|
| 16 |
+
%\yqe{
|
| 17 |
+
Figure \ref{fig:pipeline} shows the two-stage segmentation pipeline. It consists of a coarse and a fine module, which share an encoder $E$ and have their own decoder $D_C$ and $D_F$. Intuitively, the coarse module aims to capture the global contextual information from the entire image and highlight the image regions possibly containing wires. Conditioned on the predictions from the coarse module, the fine module achieves high-resolution wire segmentation by only looking at local patches likely containing wires.
|
| 18 |
+
|
| 19 |
+
Given a high-resolution image $I_\textrm{glo}$, we first bilinearly downsample it to $I_\textrm{glo}^{ds}$ with a fixed size $p\times p$ and feed it into the coarse module. The module predicts the global probability map $P_\textrm{glo} = \textrm{SoftMax}(D_C(E(I_\textrm{glo}^{ds})))$ containing the activation of the wire regions.% with rich global contextual information.
|
| 20 |
+
|
| 21 |
+
For each patch $I_\textrm{loc}$ of size $p \times p$ cropped from the full-resolution image $I_\textrm{glo}$, and the corresponding conditional probability map $P_\textrm{con}$ cropped from $P_\textrm{glo}$, we predict the local probability $P_\textrm{loc} = \textrm{SoftMax}(D_F(E(I_\textrm{loc}, P_\textrm{con})))$.
|
| 22 |
+
%$M$ is a binary location mask indicating the patch region of $I_{loc}$ in $I_{glo}$ with 1, and elsewhere 0.
|
| 23 |
+
Note that $E$ is shared between the coarse and the fine module, thus it should take inputs with the same number of channels. Therefore, for the coarse module, we concatenate an additional zero channel with the input image to make the channel number consistent.
|
| 24 |
+
% We formulate our problem as a semantic segmentation task, so we can regard it as a pixel-wise classification problem. Each image pixel is labeled as 0 or 1 indicating background or wire pixel.
|
| 25 |
+
|
| 26 |
+
% We apply Cross Entropy (CE) loss to both the global $P_\textrm{glo} = \textrm{SoftMax}(Z_\textrm{glo})$ and local probability map $P_\textrm{loc} = \textrm{SoftMax}(Z_\textrm{loc})$, comparing with their ground truth annotations $G_\textrm{glo}$ and $G_\textrm{loc}$.
|
| 27 |
+
|
| 28 |
+
We apply Cross Entropy (CE) loss to both the global $P_\textrm{glo}$ and local probability map $P_\textrm{loc}$, comparing with their ground truth annotations $G_\textrm{glo}$ and $G_\textrm{loc}$.
|
| 29 |
+
|
| 30 |
+
\vspace{-2mm}
|
| 31 |
+
\begin{equation}
|
| 32 |
+
\begin{aligned}
|
| 33 |
+
% \mathcal{L}_{glo} &= -\sum_{c\in C}\log\ Softmax(Z_{glo})_c \\
|
| 34 |
+
%\mathcal{L}_{loc} &= -\sum_{c\in C}\log\ Softmax(Z_{loc})_c \\
|
| 35 |
+
\mathcal{L}_\textrm{glo} &= CE (P_\textrm{glo}, G_\textrm{glo}) \\
|
| 36 |
+
\mathcal{L}_\textrm{loc} &= CE (P_\textrm{loc}, G_\textrm{loc}) \\
|
| 37 |
+
\end{aligned}
|
| 38 |
+
\end{equation}
|
| 39 |
+
%Both $Z_{glo}$ and $Z_{loc}$ are trained by computing Cross Entropy losses of the logit maps after softmax against their respective ground truth wire annotations.
|
| 40 |
+
The final loss $\mathcal{L}$ is the sum of the two:
|
| 41 |
+
|
| 42 |
+
\vspace{-3mm}
|
| 43 |
+
\begin{equation}
|
| 44 |
+
\begin{aligned}
|
| 45 |
+
\mathcal{L} &= \mathcal{L}_ \textrm{glo} + \lambda \mathcal{L}_ \textrm{loc},
|
| 46 |
+
\end{aligned}
|
| 47 |
+
\end{equation}
|
| 48 |
+
|
| 49 |
+
%Given a high-resolution image $I$, the coarse module of our two-stage model aims to recognize semantics of the entire image and pick up wire regions at a coarse level. Looking at the entire image allows the coarse module to capture more contextual information. To fit the large images into a GPU for training and inference, images are first bilinearly downsampled, then fed into the coarse module to predict a global logit map $Z_{glo}$. We use $Z_{glo}$ as a conditional input to guide the fine module in the next stage.
|
| 50 |
+
|
| 51 |
+
%The fine module is conditioned on the output from the coarse module. This module takes in a local image patch $I_{loc}$ of the whole image at full resolution, the global logit map $Z_{glo}$ from the coarse module, and a binary location mask $M$ that sets the patch relative to the full image to $1$ and other regions to $0$. The fine module then predicts the local wire logit map $Z_{loc}$. Empirically, we find that concatenating the entire global logit map rather than cropping the logit map at the location of the image patch yields slightly improved results.
|
| 52 |
+
|
| 53 |
+
%The designs of the coarse and fine modules are conceptually the same as those in GLNet~\cite{glnet} and MagNet~\cite{magnet}, where a global network is trained on entire downsampled images and a local network is trained on higher-resolution image patches conditioned on some global features. However, unlike GLNet, where intermediate features are shared between the global and local branch bidirectionally, we opt for a simpler late fusion by concatenating the logit map directly to the fine module. We also only use two stages instead of up to four stage as done in MagNet, since a single fine module is already sufficient at refining annotations that are only several pixels thick, and additional stages can drastically increase inference time.
|
| 54 |
+
|
| 55 |
+
where we
|
| 56 |
+
%empirically
|
| 57 |
+
set $\lambda = 1$ for training.
|
| 58 |
+
% During training,
|
| 59 |
+
% we apply data augmentation including random scaling, horizontal flipping and photometric distortion to the full-resolution image to obtain the input $I_{glo}$. After that,
|
| 60 |
+
% we randomly crop one local patch $I_{loc}$ with patch size $p=512$ from the augmented image.
|
| 61 |
+
Similar to Focal loss~\cite{focal} and Online Hard Example Mining~\cite{ohem}, we balance the wire and background samples in the training set by selecting patches that contain at least 1\% of wire pixels. %We find that this simple constrained cropping approach yields better performances than other well known balancing methods, including Focal loss~\cite{focal} and Online Hard Example Mining~\cite{ohem}.
|
| 62 |
+
|
| 63 |
+
%The global image $I_{glo}$ is generated by simply downsampling this augmented image. We then generate the local image patch $I_{loc}$ by randomly cropping a 512$\times$512 window from the augmented image that contains at least 1\% wire pixels. This helps balance between wire and background pixels. We find that this simple constrained cropping approach yields better performances than other well known balancing methods, including Focal loss~\cite{focal} and Online Hard Example Mining~\cite{ohem}.
|
| 64 |
+
|
| 65 |
+
%We share the same feature extractor network between both the coarse and fine module, but train separate feature decoders.
|
| 66 |
+
%To account for the additional inputs to the fine module, we expand the input channels of the feature extractor from 3 to 5. The last two channels are set to 0 when passing an image through the coarse module, and set to the global logit map $Z_{glo}$ and the binary location map $M$ for the fine module.
|
| 67 |
+
To perform inference, we first feed the downsampled image to the coarse module, which is the same as training.
|
| 68 |
+
% We then compute the global wire segmentation map by taking the $\mathrm{argmax}$ over the two classes for each pixel.
|
| 69 |
+
Local inference is done by running a sliding window over the entire image, where the patch is sampled only when there is at least some percentage of wire pixels (determined by~$\alpha$). This brings two advantages: First, we save computation time in regions where there are no wires. Second, the local fine module can leverage the information from the global branch for better inference quality.
|
| 70 |
+
|
| 71 |
+
% The design of our two-stage pipeline is conceptually similar to GLNet~\cite{glnet} and MagNet~\cite{magnet}, where a global network is trained on downsampled images and a local network is trained on image patches at the original resolution. However, unlike GLNet whose intermediate features are shared between the global and local branch bidirectionally, we opt for a simpler late fusion by concatenating the probability map directly to the fine module. We only use two stages instead of four stages in MagNet. Our pipeline demonstrates its efficiency and simplicity while coping with wire-like objects which are only several pixels thick, and avoids high computational demands from some previous works.
|
| 72 |
+
|
| 73 |
+
\subsection{Wire Feature Preservation}
|
| 74 |
+
\label{sec:wire_feature_preserve}
|
| 75 |
+
As wires are thin and sparse, applying downsampling to the input images may make the wire features vanish entirely. To mitigate this challenge, we propose a simple feature augmentation technique by taking the min and max pixel luminance values of the input image over a local window. Either the local min or the max value makes the wire pixels more visually apparent. In practice, we concatenate the min- and max-filtered luminance channels to the RGB image and condition map, resulting in 6 total channels as input. We name this component MinMax.
|
| 76 |
+
%to form a 5-channel input.
|
| 77 |
+
% Unlike common object segmentation, with prominent textures and colors, wires, especially outdoor cables, tend to either be shadowed or reflective under sunlight. As a result, it is common to see either very bright or very dark wires in images. An intuitive approach is to find the min and max of an image before downsampling, and preserve this feature as input to the model.
|
| 78 |
+
% \todo{show an example patch of luminance min max filter:}
|
| 79 |
+
|
| 80 |
+
%\subsection{Motivation}
|
| 81 |
+
% We design a two-stage coarse-to-fine pipeline for wire-like object segmentation. This model structure
|
| 82 |
+
% we divide our model into two major components -- a coarse module and a fine module. Our two-stage model design
|
| 83 |
+
% is motivated by two observations. First, wires in our dataset are extremely long and thin, many spanning over several thousand pixels but only occupy several pixels across. Limited by the memory size of modern GPUs, we cannot simply pass the entire image at full-resolution to a model for inference. As a result, two separate modules are required for high-resolution inference, where the coarse module captures the entire wire and its surrounding contextual information at a lower resolution, and the fine module captures detailed textures of the wire at the original resolution.
|
| 84 |
+
|
| 85 |
+
% Second, we observe that wire pixels are sparse, where a typical high-resolution image contains a small percentages of wire pixels. This means that we can also use the coarse module to guide the fine module on what regions to capture the details. The fine module can save computation by only predicting segmentation masks where there are wires.
|
| 86 |
+
|
| 87 |
+
% \subsection{Global Condition}
|
| 88 |
+
% Many high-resolution segmentation methods share the idea of using global-local refinement module. But each method differs in details that are tailored to their targeted applications.
|
| 89 |
+
% MagNet~\cite{magnet} trains a single model by randomly sampling global/local patches, and then trains a refinement module that takes as input the two predictions. This neglects the causal relationship between global and local features. The refinement module does not take into account any image feature. The recently proposed ISDNet~\cite{isdnet} use a shallow network and takes the entire image as input. This is proven effective in some high-resolution segmentation datasets such as DeepGlobe~\cite{deepglobe}, where resolution is fixed and manageable (5k$\times$5k). But the model is unable to scale to larger images. The shallow model, while being efficient, is limited in model capacity, especially with sparse labels.
|
| 90 |
+
% the fact that they take the entire image as input strictly limits their model to use lightweight networks as the shallow branch (i.e. STDC~\cite{stdc} with output stride=8), which has limited performance in cases where the target label is also small.
|
| 91 |
+
% In fact, we were unable to train ISDNet with a stronger shallow network to produce high-resolution outputs.
|
| 92 |
+
% ~\cite{learning_downsample} attempts to learn a downsampling network, which we tried but found was detrimental for our thin and sparse wire masks.
|
| 93 |
+
% As a result,
|
| 94 |
+
|
| 95 |
+
Besides feature augmentations, we also adapt the architecture to maximally preserve the sparse wire annotations.
|
| 96 |
+
% Different from existing methods, we find it crucial to keep the image at a close-to-original resolution to maximally preserve the image features of wires. In addition, we find it similarly important to preserve the sparse wire annotations.
|
| 97 |
+
% We use the global branch as a conditioning network for training both coarse and high resolution segmentation tasks rather than a simple coarse branch that is only learned for the coarse resolution.
|
| 98 |
+
We propose to use ``overprediction" and achieve this by using max-pool downsampling on the coarse labels during training, which preserves activation throughout the coarse branch. We name this component MaxPool. We provide ablation studies for these components in Section~\ref{sec:results}.
|
| 99 |
+
|
| 100 |
+
\vspace{-5mm}
|
| 101 |
+
\section{High-Resolution Wire Inpainting}
|
| 102 |
+
\label{sec:inpainting}
|
| 103 |
+
Given a full-resolution wire segmentation mask estimated by our wire segmentation model, we propose an inpainting pipeline to remove and fill in the wire regions. Our approach addresses two major challenges in wire inpainting. First, recent state-of-the-art deep inpainting methods do not handle arbitrary resolution images, which is critical for high-resolution wire removal. Second, deep inpainting methods often suffer from color inconsistency when the background has uniform (or slowly varying) colors. This issue is particularly significant for wires, as they are often in front of uniform backgrounds, such as the sky or building facades. The commonly used reconstruction loss, such as L1, is not sensitive to color inconsistency, which further exacerbates this issue.
|
| 104 |
+
|
| 105 |
+
We thus revisit the efficient deep inpainting method LaMa \cite{suvorov2022resolution}. Compared with other inpainting models, LaMa has two major advantages. First, it contains the Fourier convolutional layers which enables an efficient and high-quality structural completion. This helps complete building facades and other man-made structures with fewer artifacts. Second, its high inference efficiency makes a tile-based inference approach possible for high resolution images.
|
| 106 |
+
|
| 107 |
+
% To tailor LaMa for wire removal, we fix the input size of the model to $512\times512$ and train the model on an augmented dataset by including synthetic wire masks and cropped patches from full-resolution images.
|
| 108 |
+
To address color inconsistency, we propose a novel ``onion-peel" color adjustment module. Specifically, we compute the mean of the RGB channels within the onion-peel regions $M_o = D(M, d) - M$ of the wire mask $M$, where $D$ is the binary dilation operator, and $d$ is the kernel size. The color difference for each channel $c \in {R, G, B}$ becomes $\textrm{Bias}_c = \mathbb{E}[M_o (x_c - y_c)]$, where $x$ is the input image, and $y$ is the output from the inpainting network. The final output of the inpainting model is: $\hat{y_c} = y_c + \textrm{Bias}_{c}$. Loss functions are then applied to $\hat{y_c}$ to achieve color consistency while compositing the final result $y_{out} = (1 - M) \odot x + M \odot \hat{y}$.
|
| 109 |
+
|
| 110 |
+
% While running inference on full-resolution images, we apply a tile-based approach, by fixing the window size at $512\times 512$ with an $32$-pixel overlap.
|
| 111 |
+
% This makes the model consistent in training and testing settings, and gives good textural and structural details for local regions.
|
paper-tex/sections/related_work.tex
ADDED
|
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\section{Related Work} \label{sec:related}
|
| 3 |
+
|
| 4 |
+
\input{figure_tex/motivation}
|
| 5 |
+
|
| 6 |
+
% \textcolor{blue}{1. semantic segemtnation,however, as we will discuss, directly using these methods will nto help. , 3. high resolution: xxx, - directly using them is not working 3. wire segmentnion.}
|
| 7 |
+
|
| 8 |
+
%We introduce related works in the following way. We first describe several widely-used semantic segmentation methods, followed by methods that tackle more specific challenges such as high-resolution segmentation and wire segmentation. We describe key principles of these works, where we can draw inspirations and improve upon.
|
| 9 |
+
|
| 10 |
+
\paragraph{Semantic segmentation}
|
| 11 |
+
Semantic segmentation has been actively researched over the past decade. For example, the DeepLab series~\cite{deeplab, deeplabv3, deeplabv3p} has been one of the most widely used set of semantic segmentation methods. They leverage dilated convolutions to capture long-range pixel correlations. Similarly, CCNet~\cite{ccnet} attend to non-local regions via a two-step axial attention mechanism. PSPNet~\cite{pspnet} use multi-scale pooling to extract high-resolution features.
|
| 12 |
+
|
| 13 |
+
Recently, the self-attention mechanism~\cite{attention} has gained increasing popularity.
|
| 14 |
+
% Originally applied in Natural Language Processing tasks, its vision-based variants~\cite{vit, swin} have shown to be superior to traditional convolution-based methods.
|
| 15 |
+
Transformer-based models for semantic segmentation~\cite{dpt, setr, swin, segformer, hassani2022neighborhood, hassani2022dilated, jain2021semask, jain2022oneformer} significantly outperform convolution-based networks since the attention modules benefit from their global receptive fields~\cite{segformer}, which let the models attend to objects that span across larger portions of the feature map.
|
| 16 |
+
|
| 17 |
+
While these above methods work well in common object semantic segmentation, when applied to our task of wire segmentation in high-resolution images, they either drop significantly in segmentation quality or require long inference times. We show in Section~\ref{sec:results} that directly applying these methods to our task yields undesirable results.
|
| 18 |
+
|
| 19 |
+
\vspace{-5mm}
|
| 20 |
+
\paragraph{High-resolution image segmentation}
|
| 21 |
+
Segmentation in high-resolution images involves additional design considerations. It is computationally infeasible to perform inference on the full-resolution image with a deep network. As a result, to maximally preserve image details within the available computation resources, many methods employ a global-local inference pipeline. For instance, GLNet~\cite{glnet} simultaneously predict a coarse segmentation map on the downsampled image and a fine segmentation map on local patches at the original resolution, then fuse them to produce the final prediction.
|
| 22 |
+
%Their model shares features produced by both the global and local branch, thus achieving feature fusion.
|
| 23 |
+
MagNet~\cite{magnet} is a recent method that proposes to iteratively predict and refine coarse segmentation maps at multiple scales using a single feature extractor and multiple lightweight refinement modules. CascadePSP~\cite{cascadepsp} train a standalone class-agnostic model to refine predictions at a higher resolution from a pretrained segmentation model. ISDNet~\cite{isdnet} propose to use an extremely lightweight subnetwork to take in the entire full-resolution image. However, the subnetwork is limited in capacity and thus segmentation quality. We share the same idea with these past works on using a coarse-to-fine approach for wire segmentation, but modify the architecture and data processing to tailor to wires.
|
| 24 |
+
|
| 25 |
+
% Most of these methods share a similar coarse-to-fine approach, where instead of performing a single inference on high-resolution images, they divide them into a coarse-to-fine approach. We repurpose important components of these methods into a two-stage model design optimized for wire segmentation in high-resolution images.
|
| 26 |
+
%While our model pipeline is similar to the above methods, there are a few fundamental differences that makes our pipeline more effective and efficient in our task. First, GLNet essentially uses two separate networks for their global and local branches. The combined network does not share weights, and requires a three-stage training scheme. Meanwhile, CascadePSP only does segmentation refinement, which means they require an entire separate network to produce the initial prediction. MagNet uses an iterative refinement method at multiple scales, where the inference time scales quadratically with number of refinement scales. They also use a much smaller module for refinement, which limits quality of refinement. In contrast, our proposed network shares the feature extractor and can be trained end-to-end. Our model is also capable of predicting accurate wire masks on very high-resolution images with only two stages and without requiring any separately trained model for initial prediction.
|
| 27 |
+
\vspace{-5mm}
|
| 28 |
+
\paragraph{Wire/Curve segmentation}
|
| 29 |
+
While few works tackle wire segmentation in high-resolution images, there are prior works that handle similar objects. For example, Transmission Line Detection (TLD) is an actively researched area in aerial imagery for drone applications. Convolutional neural networks are used~\cite{ttpla, pldu, cable_inst, lsnet} to segment overhanging power cables in outdoor scenes. However, wire patterns in TLD datasets are relatively consistent in appearance and shape -- evenly spaced and only spanning locally. In contrast, we handle more generic wires seen in regular photographic contents, where the wire appearance has much higher variety.
|
| 30 |
+
%We will discuss in Section~\ref{sec:results} that the models from these works do not generalize to wires in our task.
|
| 31 |
+
|
| 32 |
+
Some other topics are loosely related to our task. Lane detection~\cite{lanedet,swiftlane,structurelane} aims to segment lanes for autonomous driving applications. These methods benefit from simple line parameterization (e.g., as two end-points), and strong positional priors. In contrast, as shown in Figure~\ref{fig:motivation}, wires vary drastically in shapes and sizes in our task, thus making them difficult to parameterize.
|
| 33 |
+
|
| 34 |
+
\vspace{-5mm}
|
| 35 |
+
\paragraph{High-Resolution Image Inpainting}
|
| 36 |
+
Image inpainting has been well-explored using patch synthesis-based methods \cite{barnes2009patchmatch, wexler2007space, darabi2012image, kaspar2015self} or deep neural networks \cite{contextencoder, globallocal, partialconv, contextual, yu2019free, xu2022image}. Zhao \textit{et al.} leveraged the powerful image sysnthesis ability of StyleGAN2 \cite{karras2020analyzing} and proposed CoModGAN \cite{comodgan} to push the image generation quality to a newer level, and was followed by \cite{zheng2022cm, jain2022keys}. Most of these deep models cannot be applied to inpainting tasks at high-resolution images. The latest diffusion-based inpainting model like DALLE-2 \cite{dalle}, LDM \cite{rombach2022high}, and StableDiffusion etc. also suffer from long inference time and low output resolution. ProFill \cite{zeng2020high} was first proposed to address high resolution inpainting via a guided super resolution module. HiFill \cite{hifill} utilized a contextual residual aggregation module and the resolution can be up to 8K. LaMa \cite{suvorov2022resolution} applied the fourier convoluational residual blocks to make the propagation of image structures well. LaMa was trained on only $256 \times 256$ images, but can be used for images up to 2K with high quality. Recently, Zhang \textit{et al.} \cite{supercaf} proposed to use guided PatchMatch for any-resolution inpainting and extended the deep inpainting results from LaMa to modern camera resolution. The textures are better reused, while the structure and line completion at high-resolution can still be challenging. In this paper, we aim at removing wires from high resolution photos. The problem can become easier if we run inpainting in a local manner since wires are usually thin and long. Therefore, we propose to revisit LaMa for wire removal, and run the inference in a tile-based fashion.
|
paper-tex/sections/results.tex
ADDED
|
@@ -0,0 +1,156 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\section{Experiments} \label{sec:results}
|
| 3 |
+
|
| 4 |
+
\subsection{Implementation Details}
|
| 5 |
+
\paragraph{Wire Segmentation Network.}
|
| 6 |
+
We experiment with ResNet-50~\cite{resnet} and MixTransformer-B2~\cite{segformer} as our shared feature extractor. We expand the input RGB channel to six channels by concatenating the conditional probability map, min- and max-filtered luminance channels. For the min and max filtering, we use a fixed 6x6 kernel. We use separate decoders for the coarse and fine modules, denoted as $D_C$ and $D_F$ respectively.
|
| 7 |
+
|
| 8 |
+
%\textcolor{purple}{
|
| 9 |
+
We use the MLP decoder proposed in~\cite{segformer} for the MixTransformer segmentation model, and the ASPP decoder in~\cite{deeplabv3p} for our ResNet-50 segmentation model. In both the segmentation and inpainting modules, we take the per-pixel average of the predicted probability when merging overlapping patches. To crop $P_\mathrm{con}$ from $P_\mathrm{glo}$, we upsample the predicted $P_\mathrm{glo}$ to the original resolution, then crop the predicted regions according to the sliding window position.
|
| 10 |
+
%}
|
| 11 |
+
|
| 12 |
+
%\textcolor{purple}{
|
| 13 |
+
To train the segmentation module, we downsample the image $I_\mathrm{glo}$ to $p\times p$ to obtain $I_\mathrm{glo}^\mathrm{ds}$. From $I_\mathrm{glo}$, we randomly crop one $p\times p$ patch $I_\mathrm{loc}$ that contains at least 1\% wire pixels. This gives a pair of $I_\mathrm{glo}^\mathrm{ds}$ and $I_\mathrm{loc}$ to compute the losses. During inference, $I_\mathrm{glo}^\mathrm{ds}$ is obtained in the same way as training, while multiple $I_\mathrm{loc}$ are obtained via a sliding window sampled only when the proportion of wire pixels is above $\alpha$. All feature extractors are pretrained on ImageNet.
|
| 14 |
+
%, while the decoders are trained from scratch.
|
| 15 |
+
%}
|
| 16 |
+
|
| 17 |
+
We train our model on 5000 training images. The model is trained for 80k iterations with a batch size of 4. We set patch size $p = 512$ during training.
|
| 18 |
+
%Global images are downsampled to 512$\times$512. We feed the global images to the coarse module and obtain a two channel class logit map. A single 1$\times$1 convolution is used to transform this logit map into a single channel map, which is then concatenated with the local RGB image and binary location mask. The five-channel input is finally fed into the fine module to obtain the local logit map.
|
| 19 |
+
For all ResNet models, we use SGD with a learning rate of 0.01, a momentum of 0.9 and weight decay of 0.0005. For MixTransformer models, we use AdamW~\cite{adamw} with a learning rate of 0.0002 and weight decay of 0.0001. Our training follows the ``poly" learning rate schedule with a power of 0.9. During inference, we set both the global image size and local patch size $p$ to 1024. Unless otherwise specified, we set the percentage for local refinement to $1\%$ ($\alpha=0.01$).
|
| 20 |
+
|
| 21 |
+
%In this section, we provide a set of quantitative results and comparisons between our two-stage model and other methods. We then provide ablation studies, qualitative visualizations and discuss several failure cases. For additional results and details, please refer to our supplementary materials.
|
| 22 |
+
|
| 23 |
+
%\todo{illurstrate the sampling strategy in training and inference strategy here?}
|
| 24 |
+
\vspace{-4mm}
|
| 25 |
+
\paragraph{Wire Inpainting Network.}
|
| 26 |
+
We adopt LaMa \cite{suvorov2022resolution} for wire inpainting by finetuning on an augmented wire dataset.
|
| 27 |
+
To prepare the wire training set, we randomly crop ten $680\times680$ patches from the non-wire regions of each image in our training partition. In total, we have 50K more training images in addition to
|
| 28 |
+
%\textcolor{purple}{
|
| 29 |
+
the 8M
|
| 30 |
+
%}
|
| 31 |
+
Places2 \cite{zhou2017places} dataset, and increase its sampling rate %\textcolor{purple}{
|
| 32 |
+
by $10\times$
|
| 33 |
+
%}
|
| 34 |
+
to balance the dataset. We also use all the ground truth segmentation maps in our training set to sample wire-like masks. During training, we start from Big-LaMa weights, and train the model on $512\times 512$ patches. We also prepare a synthetic wire inpainting quality evaluation dataset, containing 1000 images at $512\times 512$ with synthetic wire masks.
|
| 35 |
+
%containing 1000 images of $512 \times 512$ and synthetic wire masks on that.
|
| 36 |
+
While running inference on full-resolution images, we apply a tile-based approach, by fixing the window size at $512\times 512$ with an $32$-pixel overlap.
|
| 37 |
+
|
| 38 |
+
\subsection{Wire Segmentation Evaluation}
|
| 39 |
+
\paragraph{Quantitative Evaluation}
|
| 40 |
+
|
| 41 |
+
%\subsection{Globally Conditioned Inference}
|
| 42 |
+
|
| 43 |
+
We compare with several widely-used object semantic segmentation and high-resolution semantic segmentation models. Specifically, we train DeepLabv3+~\cite{deeplabv3p} with ResNet-50~\cite{resnet} backbone under two settings: global and local. In the global setting, the original images are resized to 1024$\times$1024. In the local setting, we randomly crop 1024$\times$1024 patches from the original images. We train our models on 4 Nvidia V100 GPUs and test them on a single V100 GPU. For high-resolution semantic segmentation models, we compare with CascadePSP~\cite{cascadepsp}, MagNet~\cite{magnet} and ISDNet~\cite{isdnet}. We describe the training details of these works in the supplement.
|
| 44 |
+
|
| 45 |
+
We present the results of in Table~\ref{table:results} tested on \benchmark. We report wire IoU, F1-score, precision and recall for quantitative evaluation. We also report wire IoUs for images at three scales, small (0 -- 3000$\times$3000), medium (3000$\times$3000 -- 6000$\times$6000) and large (6000$\times$6000+), which are useful for analyzing model characteristics. Finally, we report average, minimum and maximum inference times on \benchmark.
|
| 46 |
+
|
| 47 |
+
\input{tables/results}
|
| 48 |
+
|
| 49 |
+
As shown in Table~\ref{table:results}, while the global model runs fast, it has lower wire IoUs. In contrast, the local model produces high-quality predictions, but requires a very long inference time.
|
| 50 |
+
Meanwhile, although CascadePSP is a class-agnostic refinement model designed for high-resolution segmentation refinement, it primarily targets common objects and does not generalize to wires.
|
| 51 |
+
% \textcolor{red}{\sout{We thus retrain CascadePSP on our dataset but find their data perturbation does not realistic model coarse wire segmentation, thus cannot effectively conduct refinement.}}
|
| 52 |
+
For MagNet, its refinement module only takes in probability maps without image features, thus failing to refine when the input prediction is inaccurate.
|
| 53 |
+
% As a result, the refinement module cannot accurately produce high-resolution wire predictions.
|
| 54 |
+
Among these works, ISDNet is relatively effective and efficient at wire prediction.
|
| 55 |
+
% In fact, their inference time is on par with the global network even at high image resolution, while maintaining relatively high wire IoU.
|
| 56 |
+
However, their shallow network design trades off capacity for efficiency, limiting the performance of wire segmentation that is thin and sparse.
|
| 57 |
+
% \textcolor{red}{\sout{For a fair comparison, we tried to replace the shallow branch in ISDNet with a MixTransformer backbone but failed due to GPU memory limitation. We thus only replace their deep network module, which only yields minor improvement, as shown in Table~\ref{table:results}.}}
|
| 58 |
+
|
| 59 |
+
% \textcolor{purple}{
|
| 60 |
+
Compared to the methods above, our model achieves the best trade-off between accuracy and memory consumption. By leveraging the fact that wires are sparse and thin, our pipeline captures both global and local features more efficiently, thus saving a lot of computation while maintaining high segmentation quality.
|
| 61 |
+
% }
|
| 62 |
+
% There are two main reasons for this poor performance.
|
| 63 |
+
% First, the global model performs inference on downsampled images, which leads to degraded image quality around thin wire-like objects and causes imprecise or disjointed predictions. Second, since prediction maps are upsampled to the original resolution, the final segmentation map may contain artifacts such as aliasing and over-predicted regions.
|
| 64 |
+
% \mt{
|
| 65 |
+
% First, since whole-image models predict segmentation maps on downsampled images, predictions on extremely small wires may be incomplete or of low quality. Second, whole-image segmentation maps are obtained by bilinearly upsampling the model output which leads to loose predictions.
|
| 66 |
+
% }
|
| 67 |
+
% These issues are rare in common object semantic segmentation since their maximum downsampling rate is usually no greater than 3$\times$, while the downsampling rate in our task can reach 10$\times$ (e.g. downsampling 10k$\times$10k images to 1024$\times$1024.). As a result, common object semantic segmentation methods fail to maintain their performances in our task.
|
| 68 |
+
|
| 69 |
+
\input{figure_tex/comparison}
|
| 70 |
+
|
| 71 |
+
% Meanwhile, local models yield higher IoUs, but have significantly longer inference times. This is because in order to preserve resolutions of very large images, many iterations of sliding-window have to be performed (e.g. 25 iterations are required for an image with size 5120$\times$5120 and a window size of 1024$\times$1024), which leads to significantly slower inference speeds.
|
| 72 |
+
|
| 73 |
+
% In contrast, our two-stage model achieves better results than sliding-window models while only requiring less than half the inference time. The coarse module in our two-stage model determines potential wire regions for refinement, and skips the refinement step on regions with no wires. This saves inference time and increases the overall inference speed. In regions where there are wires, the fine module leverages information from the coarse module to more accurately predict a tight segmentation mask. These two factors together yield an effective and efficient model for wire semantic segmentation on high-resolution images.
|
| 74 |
+
\vspace{-4.5mm}
|
| 75 |
+
% \subsubsection{Comparing with SOTA}
|
| 76 |
+
\paragraph{Qualitative Evaluation}
|
| 77 |
+
%\todo{section subject to change}\\
|
| 78 |
+
We provide visual comparisons of segmentation models in Figure~\ref{fig:visual}. We show the ``local'' DeepLabv3+ model as it consistently outperforms its ``global'' variant given that ``local'' predicts wire masks in a sliding-window manner at the original image resolution. As a trade-off, without global context, the model suffers from over-prediction. CascadePSP is designed to refine common object masks given a coarse input mask, thus fails to produce satisfactory results when the input is inaccurate or incomplete. Similarly, the refinement module of MagNet does not handle inaccurate wire predictions. ISDNet performs the best among related methods, but the quality is still unsatisfactory as it uses a lightweight model with limited capacity. Compared to all these methods, our model captures both global context and local details, thus producing more accurate mask predictions.
|
| 79 |
+
\vspace{-4mm}
|
| 80 |
+
\paragraph{Ablation Studies}
|
| 81 |
+
|
| 82 |
+
% \subsubsection{Effectiveness of global logit map}
|
| 83 |
+
|
| 84 |
+
% We show that using the global logit map as input to the fine module conveys more contextual information effectively than using the local logit map. For comparison, we train a separate two-stage model, where we only crop and resize the logit map at the location of the local image patch as input to the fine module.
|
| 85 |
+
|
| 86 |
+
% \input{tables/logit}
|
| 87 |
+
|
| 88 |
+
% As shown in Table~\ref{table:logit}, using only the local logit map yields inferior performances. Specifically, we find that in situations where local refinement is taken place, the global logit map provides sufficient information for the fine module to identify confusing non-wire objects, thus avoiding over-prediction. Figure~\ref{fig:overpredict} demonstrates that our fine module with global logit map input successfully avoids a pattern on the building that strongly resembles a wire.
|
| 89 |
+
|
| 90 |
+
% \input{figure_tex/overpredict}
|
| 91 |
+
|
| 92 |
+
In Table~\ref{table:component_ablation}, we report wire IoUs after removing each component in our model, including MinMax, MaxPool, and Coarse condition concatenation. We find that all components play a significant role for accurate wire prediction, particularly in large images. Both MinMax and MaxPool are effective in encouraging prediction, which is shown by the drop in recall without either component, also shown in Figure~\ref{fig:visual}. Coarse condition, as described in Section~\ref{sec:segmentation}, is crucial in providing global context to the local network, without which the wire IoU drops significantly.
|
| 93 |
+
|
| 94 |
+
Table~\ref{table:thresholds} shows the wire IoUs and inference speed of our two-stage model as $\alpha$ changes. We observe a consistent decrease in performance as $\alpha$ increases. On the other hand, setting $\alpha$ to 0.01 barely decreases IoU, while significantly boosting inference speed, which means the coarse network is effectively activated at wire regions.
|
| 95 |
+
% We believe that in addition to providing a condition for refinement, the coarse module also acts as a suppressor to eliminate false positives, it does so by skipping potential false positives that would otherwise be mis-classified as wires. As a result, setting the optimal $\alpha$ allows the network to predict accurate segmentation masks, avoid over-prediction, and maintain a high inference speed all at the same time.
|
| 96 |
+
|
| 97 |
+
% Note that when $\alpha=0.0\%$ (refine on all windows), our model still outperforms a single sliding-window model (70.3\% vs. 69.0\%). This means that the coarse module indeed provides useful information to the fine module via the logit map, which further justifies our two-stage design.
|
| 98 |
+
|
| 99 |
+
% \subsubsection{Effectiveness of model components}
|
| 100 |
+
|
| 101 |
+
% We find that the global context provides vital information to the local branch during inference to avoid false positives. When the global branch softmax is fed into the local branch together with the local image, the network suppresses predictions at regions that look like wires, such as pavement cracks. We show these in Figure X. Quantitatively, the wire IoU drops significantly without conditioning on the global branch.
|
| 102 |
+
|
| 103 |
+
% \subsubsection{Effectiveness of Maxpool resize}
|
| 104 |
+
|
| 105 |
+
% \subsubsection{Effectiveness of MinMax input}
|
| 106 |
+
|
| 107 |
+
% Here we show the useful information from the minmax image. Figure X compares a model trained with/without minmax. As can be seen, for extremely bright/dark wires, minmax is able to emphasize this feature and ensure prediction in those areas where they are not easily seen after downsampling.
|
| 108 |
+
|
| 109 |
+
|
| 110 |
+
|
| 111 |
+
\subsection{Wire Inpainting Evaluation}
|
| 112 |
+
\vspace{-1mm}
|
| 113 |
+
We evaluate our wire inpainting model using the synthetic dataset. Results are shown in Table~\ref{exp:wire_inp}. Our model structure is highly related to LaMa~\cite{suvorov2022resolution}. The difference is the training data and the proposed color adjustment module to address color inconsistency. We also compare our methods with PatchMatch \cite{barnes2009patchmatch} based on patch synthesis, DeepFillv2~\cite{yu2019free} based on Contextual Attention, CMGAN~\cite{zheng2022cm} and FcF~\cite{jain2022keys} based on StyleGAN2~\cite{karras2020analyzing} and LDM~\cite{rombach2022high} based on Diffusion. Inference speed is measured on a single A100-80G GPU. Visual results on synthetic and real images are shown in Figure \ref{fig:wire_inp}. PatchMatch, as a traditional patch synthesis method, produces consistent color and texture that leads to high PSNR. However, it performs severely worse on complicated structural completion. StyleGAN-based CMGAN and FcF are both too heavy for wires that are thin and sparse. Besides, diffusion-based models like LDM tends to generate arbitrary objects and patterns. DeepFill and the official Big-LaMa both have severe color inconsistency issue, especially in the sky region. Our model has a balanced quality and efficiency, and performs well on structural completion and color consistency.
|
| 114 |
+
Note that we use a tile-based method at inference time.
|
| 115 |
+
% with a window size of $512 \times 512$ and an overlap of $32$.
|
| 116 |
+
The reason the tile-based strategy can be employed is due to the wire characteristics: sparse, thin and lengthy. More high-resolution inpainting results are in the supplementary materials.
|
| 117 |
+
|
| 118 |
+
% \begin{figure}[h!]
|
| 119 |
+
% \centering
|
| 120 |
+
% \captionsetup{type=figure}
|
| 121 |
+
% \includegraphics[width=1.\linewidth]{figures/inpainting_result.pdf}
|
| 122 |
+
% \vspace{-6mm}
|
| 123 |
+
% \captionof{figure}{\textbf{Inpainting Comparison}. Our model performs well on complicated structure completion and color consistency, especially on building facades and sky regions containing plain and uniform color.
|
| 124 |
+
% \vspace{-3mm}
|
| 125 |
+
% }
|
| 126 |
+
% \label{fig:wire_inp}
|
| 127 |
+
% \end{figure}
|
| 128 |
+
|
| 129 |
+
|
| 130 |
+
% \begin{table}[t]\setlength{\tabcolsep}{5pt}
|
| 131 |
+
% \setlength{\abovecaptionskip}{8pt}
|
| 132 |
+
% \centering
|
| 133 |
+
% \footnotesize
|
| 134 |
+
% % \scriptsize
|
| 135 |
+
% % \tiny
|
| 136 |
+
|
| 137 |
+
% %\vspace{-2ex}
|
| 138 |
+
% % \resizebox{\columnwidth}{!}{
|
| 139 |
+
% \begin{tabular}{r|c c c|c}
|
| 140 |
+
% \hline
|
| 141 |
+
% Model &PSNR$\uparrow$&LPIPS$\downarrow$&FID$\downarrow$ &Speed (s/img)\\ \hline
|
| 142 |
+
% %Photoshop\\ %should be easy to run batch testing
|
| 143 |
+
% PatchMatch \cite{barnes2009patchmatch}&50.29 &0.0294 & 5.0403 & -\\
|
| 144 |
+
% DeepFillv2 \cite{yu2019free} &47.01 &0.0374&8.0086 &0.009\\
|
| 145 |
+
% CMGAN \cite{zheng2022cm} &50.07 &0.0255 &3.8286 &0.141\\
|
| 146 |
+
% FcF \cite{jain2022keys}&48.82&0.0322&4.7848&0.048\\
|
| 147 |
+
% LDM \cite{rombach2022high} & 45.96 & 0.0401& 10.1687 & 4.280\\
|
| 148 |
+
% Big-LaMa \cite{suvorov2022resolution} & 49.63 & 0.0267& 4.1245 &0.034\\
|
| 149 |
+
% Ours (LaMa-Wire) & 50.06 & 0.0259 & 3.6950 &0.034\\
|
| 150 |
+
% \hline
|
| 151 |
+
% \end{tabular}
|
| 152 |
+
% \caption{Quantitative results of inpainting on our synthetic wire inpainting evaluation dataset (1000 images). Our model achieves the highest perceptual quality in terms of FID, and has a balanced speed and quality.}
|
| 153 |
+
% % }
|
| 154 |
+
% \label{exp:wire_inp}
|
| 155 |
+
% % \vspace{-4mm}
|
| 156 |
+
% \end{table}
|
paper-tex/supplement.bbl
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
\begin{thebibliography}{1}\itemsep=-1pt
|
| 2 |
+
|
| 3 |
+
\bibitem{cascadepsp}
|
| 4 |
+
Ho~Kei Cheng, Jihoon Chung, Yu-Wing Tai, and Chi-Keung Tang.
|
| 5 |
+
\newblock Cascadepsp: toward class-agnostic and very high-resolution
|
| 6 |
+
segmentation via global and local refinement.
|
| 7 |
+
\newblock In {\em Proceedings of the IEEE/CVF Conference on Computer Vision and
|
| 8 |
+
Pattern Recognition}, pages 8890--8899, 2020.
|
| 9 |
+
|
| 10 |
+
\bibitem{isdnet}
|
| 11 |
+
Shaohua Guo, Liang Liu, Zhenye Gan, Yabiao Wang, Wuhao Zhang, Chengjie Wang,
|
| 12 |
+
Guannan Jiang, Wei Zhang, Ran Yi, Lizhuang Ma, et~al.
|
| 13 |
+
\newblock Isdnet: Integrating shallow and deep networks for efficient
|
| 14 |
+
ultra-high resolution segmentation.
|
| 15 |
+
\newblock In {\em Proceedings of the IEEE/CVF Conference on Computer Vision and
|
| 16 |
+
Pattern Recognition}, pages 4361--4370, 2022.
|
| 17 |
+
|
| 18 |
+
\bibitem{magnet}
|
| 19 |
+
Chuong Huynh, Anh~Tuan Tran, Khoa Luu, and Minh Hoai.
|
| 20 |
+
\newblock Progressive semantic segmentation.
|
| 21 |
+
\newblock In {\em Proceedings of the IEEE/CVF Conference on Computer Vision and
|
| 22 |
+
Pattern Recognition}, pages 16755--16764, 2021.
|
| 23 |
+
|
| 24 |
+
\end{thebibliography}
|
paper-tex/supplement.tex
ADDED
|
@@ -0,0 +1,184 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
% CVPR 2023 Paper Template
|
| 2 |
+
% based on the CVPR template provided by Ming-Ming Cheng (https://github.com/MCG-NKU/CVPR_Template)
|
| 3 |
+
% modified and extended by Stefan Roth ([email protected])
|
| 4 |
+
|
| 5 |
+
\documentclass[10pt,twocolumn,letterpaper]{article}
|
| 6 |
+
|
| 7 |
+
%%%%%%%%% PAPER TYPE - PLEASE UPDATE FOR FINAL VERSION
|
| 8 |
+
% \usepackage[review]{cvpr} % To produce the REVIEW version
|
| 9 |
+
\usepackage{cvpr} % To produce the CAMERA-READY version
|
| 10 |
+
%\usepackage[pagenumbers]{cvpr} % To force page numbers, e.g. for an arXiv version
|
| 11 |
+
|
| 12 |
+
\usepackage[accsupp]{axessibility} % Improves PDF readability for those with disabilities.
|
| 13 |
+
|
| 14 |
+
% Include other packages here, before hyperref.
|
| 15 |
+
\usepackage[normalem]{ulem}
|
| 16 |
+
\usepackage{graphicx}
|
| 17 |
+
\usepackage{amsmath}
|
| 18 |
+
\usepackage{amssymb}
|
| 19 |
+
\usepackage{booktabs}
|
| 20 |
+
\usepackage{xcolor}
|
| 21 |
+
\usepackage{comment}
|
| 22 |
+
\usepackage{enumitem}
|
| 23 |
+
\usepackage{multirow}
|
| 24 |
+
\usepackage{footnote}
|
| 25 |
+
\newcommand{\todo}[1]{{\color{red}#1}}
|
| 26 |
+
|
| 27 |
+
\newcommand{\benchmark}{WireSegHR}
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
\makeatletter
|
| 31 |
+
\newcommand\footnoteref[1]{\protected@xdef\@thefnmark{\ref{#1}}\@footnotemark}
|
| 32 |
+
\makeatother
|
| 33 |
+
|
| 34 |
+
% It is strongly recommended to use hyperref, especially for the review version.
|
| 35 |
+
% hyperref with option pagebackref eases the reviewers' job.
|
| 36 |
+
% Please disable hyperref *only* if you encounter grave issues, e.g. with the
|
| 37 |
+
% file validation for the camera-ready version.
|
| 38 |
+
%
|
| 39 |
+
% If you comment hyperref and then uncomment it, you should delete
|
| 40 |
+
% ReviewTempalte.aux before re-running LaTeX.
|
| 41 |
+
% (Or just hit 'q' on the first LaTeX run, let it finish, and you
|
| 42 |
+
% should be clear).
|
| 43 |
+
\usepackage[pagebackref,breaklinks,colorlinks]{hyperref}
|
| 44 |
+
|
| 45 |
+
|
| 46 |
+
% Support for easy cross-referencing
|
| 47 |
+
\usepackage[capitalize]{cleveref}
|
| 48 |
+
\crefname{section}{Sec.}{Secs.}
|
| 49 |
+
\Crefname{section}{Section}{Sections}
|
| 50 |
+
\Crefname{table}{Table}{Tables}
|
| 51 |
+
\crefname{table}{Tab.}{Tabs.}
|
| 52 |
+
|
| 53 |
+
|
| 54 |
+
%%%%%%%%% PAPER ID - PLEASE UPDATE
|
| 55 |
+
\def\cvprPaperID{699} % *** Enter the CVPR Paper ID here
|
| 56 |
+
\def\confName{CVPR}
|
| 57 |
+
\def\confYear{2023}
|
| 58 |
+
|
| 59 |
+
|
| 60 |
+
\makeatletter
|
| 61 |
+
\def\@maketitle
|
| 62 |
+
{
|
| 63 |
+
\newpage
|
| 64 |
+
\null
|
| 65 |
+
\iftoggle{cvprrebuttal}{\vspace*{-.3in}}{\vskip .375in}
|
| 66 |
+
\begin{center}
|
| 67 |
+
% smaller title font only for rebuttal
|
| 68 |
+
\iftoggle{cvprrebuttal}{{\large \bf \@title \par}}{{\Large \bf \@title \par}}
|
| 69 |
+
% additional two empty lines at the end of the title
|
| 70 |
+
\iftoggle{cvprrebuttal}{\vspace*{-22pt}}{\vspace*{24pt}}
|
| 71 |
+
{
|
| 72 |
+
\large
|
| 73 |
+
\lineskip .5em
|
| 74 |
+
\begin{tabular}[t]{c}
|
| 75 |
+
\iftoggle{cvprfinal}{
|
| 76 |
+
\@author
|
| 77 |
+
}{
|
| 78 |
+
\iftoggle{cvprrebuttal}{}{
|
| 79 |
+
Anonymous \confName~submission\\
|
| 80 |
+
\vspace*{1pt}\\
|
| 81 |
+
Paper ID \cvprPaperID
|
| 82 |
+
}
|
| 83 |
+
}
|
| 84 |
+
\end{tabular}
|
| 85 |
+
\par
|
| 86 |
+
}
|
| 87 |
+
% additional small space at the end of the author name
|
| 88 |
+
% additional empty line at the end of the title block
|
| 89 |
+
\vspace*{-10mm}
|
| 90 |
+
\vspace{-5mm}
|
| 91 |
+
\end{center}
|
| 92 |
+
}
|
| 93 |
+
\makeatother
|
| 94 |
+
|
| 95 |
+
\title{Supplementary Material: Automatic High Resolution Wire Segmentation and Removal}
|
| 96 |
+
|
| 97 |
+
\begin{document}
|
| 98 |
+
|
| 99 |
+
%%%%%%%%% TITLE - PLEASE UPDATE
|
| 100 |
+
% \title{Automatic Wire Segmentation for Inpainting}
|
| 101 |
+
% \title{Segmentation to the Extreme: \\A Large-Scale Wire Segmentation Dataset and a Pilot Study}
|
| 102 |
+
%\title{Segmentation to the Extreme: \\A Large-Scale High-Resolution Wire Segmentation Dataset and a Pilot Study}
|
| 103 |
+
%\title{WINE: WIre NEver Appeared in Your Photos}
|
| 104 |
+
|
| 105 |
+
\maketitle
|
| 106 |
+
\thispagestyle{empty}
|
| 107 |
+
|
| 108 |
+
|
| 109 |
+
\vspace{-1mm}
|
| 110 |
+
\section{Comparison with Pixel 6}
|
| 111 |
+
\vspace{-1mm}
|
| 112 |
+
We show a visual comparison between our model and Pixel 6's ``Magic Eraser'' feature in Figure~\ref{fig:pixel6}. Without manual intervention, Google Pixel 6's ``Magic Eraser'' performs well on wires with clean background, but suffers from thin wires that are hardly visible ((A) upper), and also on wires with complicated background ((A) lower). We also pass our segmentation mask to our wire inpainting model to acquire the wire removal result, as shown in the lower image of (B).
|
| 113 |
+
|
| 114 |
+
\vspace{-1mm}
|
| 115 |
+
\section{Failure cases}
|
| 116 |
+
\vspace{-1mm}
|
| 117 |
+
We show some challenging cases where our model fails to predict accurate wire masks in Figure~\ref{fig:new_failures}. These include regions that are very similar to wires (top row), severe background blending (middle row) and extreme lighting conditions (bottom row).
|
| 118 |
+
|
| 119 |
+
\vspace{-1mm}
|
| 120 |
+
\section{Panorama}
|
| 121 |
+
\vspace{-1mm}
|
| 122 |
+
Our two-stage model leverages the sparsity of wires in natural images, and efficiently generalizes to ultra-high resolution images such as panoramas. We show one panoramic image of $11$K by $1.5$K resolution in Figure~\ref{fig:new_panorama}. Note that our method produces high-quality wire segmentation that covers wires that are almost invisible. As a result, our proposed wire removal step can effectively remove these regions.
|
| 123 |
+
|
| 124 |
+
|
| 125 |
+
\input{figure_tex/pixel6.tex}
|
| 126 |
+
\input{figure_tex/new_failure_cases.tex}
|
| 127 |
+
\input{figure_tex/new_panorama.tex}
|
| 128 |
+
|
| 129 |
+
|
| 130 |
+
\section{Segmentation and inpainting visualizations}
|
| 131 |
+
\vspace{-1mm}
|
| 132 |
+
We show our wire segmentation and inpainting results in several common photography scenes as well as in some challenging cases in Figure~\ref{fig:additional_visualizations}. We provide more visualizations of wire segmentation and subsequent inpainting results. Our model successfully handles numerous challenging scenarios, including strong backlit (top row), complex background texture (2nd row), low light (3rd row), and barely visible wires (4th row). A typical use case is shown in the last row.
|
| 133 |
+
%We also provide 20 additional samples in the attached HTML \mbox{(\textit{additional\_visualizations.html})}.
|
| 134 |
+
|
| 135 |
+
\section{Experiments on other datasets}
|
| 136 |
+
\vspace{-1mm}
|
| 137 |
+
Most existing wire-like datasets either are at low resolutions or are for specific purposes (e.g., aerial imaging) and thus do not contain the scene diversity like WireSegHR does. The suggested TTPLA~[2] dataset shares the Power Lines class with our dataset, although it only contains aerial images. Table~\ref{ttpla_exp} shows evaluation results of the TTPLA dataset on our model and also our WireSegHR dataset on the TTPLA model.
|
| 138 |
+
% We provide a model performance comparison on this dataset using two experiments.
|
| 139 |
+
% To abide by the request to refrain from significant additional experiments,
|
| 140 |
+
% We first test our trained model on the TTPLA test set. We then test the trained TTPLA model on our WireSegHR-500 test set:
|
| 141 |
+
|
| 142 |
+
|
| 143 |
+
\begin{table}[h!]
|
| 144 |
+
\resizebox{\linewidth}{!}{
|
| 145 |
+
\centering
|
| 146 |
+
\begin{tabular}{c|c|c}
|
| 147 |
+
Dataset & Model & IoU (\%) \\\hline\hline
|
| 148 |
+
\multirow{3}{*}{TTPLA (Power Line only)} & TTPLA (ResNet-50, 700$\times$ 700) & 18.9 \\
|
| 149 |
+
& Ours (ResNet-50) & 33.1 \\
|
| 150 |
+
& Ours (MiT-b2) & 42.7 \\\hline
|
| 151 |
+
\multirow{3}{*}{WireSegHR} & TTPLA (ResNet-50, 700$\times$ 700) & 3.5 \\
|
| 152 |
+
& Ours (ResNet-50) & 47.8 \\
|
| 153 |
+
& Ours (MiT-b2) & 60.8\\\hline
|
| 154 |
+
\end{tabular}}
|
| 155 |
+
\caption{Comparison with TTPLA.}
|
| 156 |
+
\vspace{-5mm}
|
| 157 |
+
\label{ttpla_exp}
|
| 158 |
+
\end{table}
|
| 159 |
+
|
| 160 |
+
% With our ResNet-50 model, MiT-b2 model, we obtain 33.1\% and 42.7\% wire IoU respectively, against 18.9\% from their original model.
|
| 161 |
+
TTPLA is trained on fixed resolution (700 $\times$ 700) and takes in the entire image for inference, which requires significant downsampling of our test set.
|
| 162 |
+
% may be fine for large structures such as transmission towers, but
|
| 163 |
+
As a result, the quality of thin wires deteriorates in both the image and the label. Our model drops in performance on the TTPLA dataset due to different annotation definitions: we annotate all wire-like objects while TTPLA only annotates power lines.
|
| 164 |
+
|
| 165 |
+
\vspace{-1mm}
|
| 166 |
+
\section{Additional training details}
|
| 167 |
+
\vspace{-1mm}
|
| 168 |
+
\paragraph{CascadePSP~\cite{cascadepsp}}
|
| 169 |
+
We follow the default training steps provided by the CascadePSP code\footnote{\label{note1}\href{https://github.com/hkchengrex/CascadePSP}{https://github.com/hkchengrex/CascadePSP}}. During training, we sample patches in the image that contain at least 1\% of wire pixels. During inference, we feed the predictions of the global DeepLabv3+ to the pretrained/retrained CascadePSP model to get the refined wire mask. In both cases, we follow the default inference code\footnoteref{note1} to obtain the final mask.
|
| 170 |
+
\vspace{-3mm}
|
| 171 |
+
\paragraph{MagNet~\cite{magnet}} MagNet\footnote{\href{https://github.com/VinAIResearch/MagNet}{https://github.com/VinAIResearch/MagNet}} obtains the initial mask predictions from a single backbone trained on all refinement scales. For a fair comparison, we adopt a 2-scale setting of MagNet, similar to our two-stage model, where the image is downsampled to $1024\times 1024$ in the global scale, and is kept at the original resolution in the local scale. To this end, we train a single DeepLabv3+ model by either downsampling the sample image to $1024\times 1024$ or randomly cropping $1024\times 1024$ patches at the original resolution. The sampled patches contain at least 1\% of wire pixels. We then train the refinement module based on the predictions from the DeepLabv3+ model, following the default setting. Inference is kept the same as the original MagNet model.
|
| 172 |
+
|
| 173 |
+
\vspace{-2mm}
|
| 174 |
+
\paragraph{ISDNet~\cite{isdnet}}
|
| 175 |
+
ISDNet\footnote{\href{https://github.com/cedricgsh/ISDNet}{https://github.com/cedricgsh/ISDNet}} performs inference on the entire image without sliding window. As a result, during training, we resize all images to $5000\times 5000$ and randomly crop $2500\times 2500$ windows, such that the input images can fit inside the GPUs. Sampled patches should contain 1\% wire pixels. During inference, all images are resized to $5000\times 5000$. We observe that this yields better results than if we keep images below $5000\times 5000$ at their original sizes.
|
| 176 |
+
|
| 177 |
+
\input{figure_tex/additional_visualizations.tex}
|
| 178 |
+
|
| 179 |
+
{\small
|
| 180 |
+
\bibliographystyle{ieee_fullname}
|
| 181 |
+
\bibliography{egbib}
|
| 182 |
+
}
|
| 183 |
+
|
| 184 |
+
\end{document}
|
paper-tex/tables/component.tex
ADDED
|
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
|
| 3 |
+
\begin{table}[t!]
|
| 4 |
+
\centering
|
| 5 |
+
\resizebox{0.49\textwidth}{!}{
|
| 6 |
+
\renewcommand{\arraystretch}{1}
|
| 7 |
+
\addtolength{\tabcolsep}{-2pt}
|
| 8 |
+
\begin{tabular}{r|c|ccc|ccc}
|
| 9 |
+
\hline
|
| 10 |
+
Model & \begin{tabular}[x]{@{}c@{}}Wire\\IoU\end{tabular} & F1 & Precision & Recall & \begin{tabular}[x]{@{}c@{}}IoU\\(Small)\end{tabular} & \begin{tabular}[x]{@{}c@{}}IoU\\(Medium)\end{tabular} & \begin{tabular}[x]{@{}c@{}}IoU\\(Large)\end{tabular} \\\hline\hline
|
| 11 |
+
%& \begin{tabular}[x]{@{}c@{}}Speed\\(s/img)\end{tabular} & \begin{tabular}[x]{@{}c@{}}Memory\\(GB)\end{tabular}\\ \hline\hline
|
| 12 |
+
Ours & 60.83 & 75.65 & 83.62 & 69.06 & 63.52 & 59.83 & 62.93 \\ \hline
|
| 13 |
+
-- MinMax & 60.01 & 75.01 & 84.87 & 67.2 & 63.67 & 58.99 & 61.97 \\
|
| 14 |
+
-- MaxPool & 59.86 & 74.89 & 85.25 & 66.78 & 61.45 & 59.40 & 60.76 \\
|
| 15 |
+
-- Coarse & 56.92 & 72.55 & 82.91 & 64.49 & 62.83 & 57.42 & 54.47 \\ \hline
|
| 16 |
+
\end{tabular}
|
| 17 |
+
\addtolength{\tabcolsep}{2pt}
|
| 18 |
+
}
|
| 19 |
+
\vspace{-1mm}
|
| 20 |
+
\caption{Ablation study of our model components.
|
| 21 |
+
% \yq{This table can be made one-column to save space.}\cezhang{+1, we're running out of space.}
|
| 22 |
+
}
|
| 23 |
+
\label{table:component_ablation}
|
| 24 |
+
\vspace{-2mm}
|
| 25 |
+
\end{table}
|
paper-tex/tables/inpaint.tex
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
|
| 3 |
+
\begin{table}[t]\setlength{\tabcolsep}{5pt}
|
| 4 |
+
\setlength{\abovecaptionskip}{8pt}
|
| 5 |
+
\centering
|
| 6 |
+
\footnotesize
|
| 7 |
+
% \scriptsize
|
| 8 |
+
% \tiny
|
| 9 |
+
|
| 10 |
+
%\vspace{-2ex}
|
| 11 |
+
% \resizebox{\columnwidth}{!}{
|
| 12 |
+
\begin{tabular}{r|c c c|c}
|
| 13 |
+
\hline
|
| 14 |
+
Model &PSNR$\uparrow$&LPIPS$\downarrow$&FID$\downarrow$ &Speed (s/img)\\ \hline
|
| 15 |
+
%Photoshop\\ %should be easy to run batch testing
|
| 16 |
+
PatchMatch \cite{barnes2009patchmatch}&50.29 &0.0294 & 5.0403 & -\\
|
| 17 |
+
DeepFillv2 \cite{yu2019free} &47.01 &0.0374&8.0086 &0.009\\
|
| 18 |
+
CMGAN \cite{zheng2022cm} &50.07 &0.0255 &3.8286 &0.141\\
|
| 19 |
+
FcF \cite{jain2022keys}&48.82&0.0322&4.7848&0.048\\
|
| 20 |
+
LDM \cite{rombach2022high} & 45.96 & 0.0401& 10.1687 & 4.280\\
|
| 21 |
+
Big-LaMa \cite{suvorov2022resolution} & 49.63 & 0.0267& 4.1245 &0.034\\
|
| 22 |
+
Ours (LaMa-Wire) & 50.06 & 0.0259 & 3.6950 &0.034\\
|
| 23 |
+
\hline
|
| 24 |
+
\end{tabular}
|
| 25 |
+
\vspace{-1mm}
|
| 26 |
+
\caption{Quantitative results of inpainting on our synthetic wire inpainting evaluation dataset (1000 images). Our model achieves the highest perceptual quality in terms of FID, and has a balanced speed and quality.}
|
| 27 |
+
% }
|
| 28 |
+
\label{exp:wire_inp}
|
| 29 |
+
% \vspace{-4mm}
|
| 30 |
+
\end{table}
|
paper-tex/tables/logit.tex
ADDED
|
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\begin{table}[h!]
|
| 3 |
+
\centering
|
| 4 |
+
\resizebox{0.98\linewidth}{!}{
|
| 5 |
+
\renewcommand{\arraystretch}{1.1}
|
| 6 |
+
\begin{tabular}{p{0.27\linewidth}|c|ccc}
|
| 7 |
+
\hline
|
| 8 |
+
Model & IoU (\%) & F1 (\%) & Precision (\%) & Recall (\%)\\ \hline\hline
|
| 9 |
+
Local logit only & 71.3 & 83.2 & 84.5 & \textbf{81.9} \\ \hline
|
| 10 |
+
Global logit + binary location map (ours) & \textbf{71.9} & \textbf{83.6} & \textbf{86.2} & 81.3\\ \hline
|
| 11 |
+
|
| 12 |
+
\end{tabular}
|
| 13 |
+
}
|
| 14 |
+
\vspace{-1mm}
|
| 15 |
+
\caption{Comparison between using only local logit map and global logit map as fine module input. By including the logit map of the entire image, our model avoids over-prediction. }
|
| 16 |
+
\label{table:logit}
|
| 17 |
+
\vspace{-2mm}
|
| 18 |
+
\end{table}
|
paper-tex/tables/results.tex
ADDED
|
@@ -0,0 +1,53 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\begin{table*}[t!]
|
| 3 |
+
\vspace{2mm}
|
| 4 |
+
\centering
|
| 5 |
+
\resizebox{0.98\textwidth}{!}{
|
| 6 |
+
\renewcommand{\arraystretch}{1.1}
|
| 7 |
+
\addtolength{\tabcolsep}{-2pt}
|
| 8 |
+
\begin{tabular}{r|c|ccc|ccc|ccc}
|
| 9 |
+
\hline
|
| 10 |
+
Model & \begin{tabular}[x]{@{}c@{}}Wire\\IoU\end{tabular} & F1 & Precision & Recall & \begin{tabular}[x]{@{}c@{}}IoU\\(Small)\end{tabular} & \begin{tabular}[x]{@{}c@{}}IoU\\(Medium)\end{tabular} & \begin{tabular}[x]{@{}c@{}}IoU\\(Large)\end{tabular} & \begin{tabular}[x]{@{}c@{}}Avg. Time\\(s/img)\end{tabular} & \begin{tabular}[x]{@{}c@{}}Min. Time\\(s/img)\end{tabular} & \begin{tabular}[x]{@{}c@{}}Max. Time\\(s/img)\end{tabular}\\ \hline\hline
|
| 11 |
+
DeepLabv3+ (Global)~\cite{deeplabv3p} & 37.77 & 54.83 & 69.68 & 45.20 & 51.62 & 38.89 & 31.89 & 0.22 & 0.07 & 0.78 \\
|
| 12 |
+
DeepLabv3+ (Local)~\cite{deeplabv3p} & 48.66 & 65.46 & 68.13 & 63.0 & 60.23 & 51.44 & 40.17 & 3.27 & 0.05 & 16.59 \\ \hline
|
| 13 |
+
CascadePSP (Pretrained)~\cite{cascadepsp} & 20.44 & 33.94 & 62.19 & 23.34 & 33.64 & 21.80 & 13.78 & 2.32 & 0.37 & 36.79 \\
|
| 14 |
+
CascadePSP (Retrained)~\cite{cascadepsp} & 26.85 & 42.33 & 52.44 & 35.49 & 48.22 & 28.97 & 15.80 & 2.25 & 0.37 & 25.37 \\
|
| 15 |
+
MagNet~\cite{magnet} & 33.71 & 50.42 & 87.69 & 35.38 & 43.59 & 32.67 & 34.48 & 3.89 & 0.54 & 17.97\\
|
| 16 |
+
MagNet-Fast~\cite{magnet} & 37.87 & 54.94 & 67.98 & 46.09 & 46.75 & 35.88 & 41.42 & 1.36 & 0.55 & 5.33 \\
|
| 17 |
+
ISDNet (R-18)~\cite{isdnet} & 46.52 & 63.50 & 77.56 & 53.75 & 55.09 & 47.15 & 43.34 & 0.29 & 0.12 & 0.86\\
|
| 18 |
+
ISDNet (MiT-b2)~\cite{isdnet} & 47.90 & 64.77 & 77.38 & 55.70 & 54.48 & 46.77 & 49.51 & 0.26 & 0.13 & 1.02 \\ \hline
|
| 19 |
+
Ours (R-50) & 47.75 & 64.64 & 74.86 & 56.87 & 60.68 & 50.19 & 38.19 & 1.24 & 0.13 & 4.67 \\
|
| 20 |
+
Ours (MiT-b2) & 60.83 & 75.65 & 83.62 & 69.06 & 63.52 & 59.83 & 62.93 & 0.82 & 0.07 & 3.36 \\ \hline
|
| 21 |
+
\end{tabular}
|
| 22 |
+
\addtolength{\tabcolsep}{2pt}
|
| 23 |
+
}
|
| 24 |
+
\vspace{-1mm}
|
| 25 |
+
\caption{Performances of common semantic segmentation and recent high-resolution semantic segmentation models on our dataset. We find that our dataset poses many challenges that high-resolution segmentation models fail to tackle effectively.
|
| 26 |
+
% \yq{This table can be made one-column to save space.}\cezhang{+1, we're running out of space.}
|
| 27 |
+
}
|
| 28 |
+
\label{table:results}
|
| 29 |
+
\end{table*}
|
| 30 |
+
|
| 31 |
+
% \begin{table}[t!]
|
| 32 |
+
% \centering
|
| 33 |
+
% \resizebox{0.49\textwidth}{!}{
|
| 34 |
+
% \renewcommand{\arraystretch}{1}
|
| 35 |
+
% \addtolength{\tabcolsep}{-2pt}
|
| 36 |
+
% \begin{tabular}{r|c|ccc|ccc}
|
| 37 |
+
% \hline
|
| 38 |
+
% Model & \begin{tabular}[x]{@{}c@{}}Wire\\IoU\end{tabular} & F1 & Precision & Recall & \begin{tabular}[x]{@{}c@{}}IoU\\(Small)\end{tabular} & \begin{tabular}[x]{@{}c@{}}IoU\\(Medium)\end{tabular} & \begin{tabular}[x]{@{}c@{}}IoU\\(Large)\end{tabular} \\\hline\hline
|
| 39 |
+
% %& \begin{tabular}[x]{@{}c@{}}Speed\\(s/img)\end{tabular} & \begin{tabular}[x]{@{}c@{}}Memory\\(GB)\end{tabular}\\ \hline\hline
|
| 40 |
+
% Ours & 60.83 & 75.65 & 83.62 & 69.06 & 63.52 & 59.83 & 62.93 \\ \hline
|
| 41 |
+
% -- MinMax & 60.01 & 75.01 & 84.87 & 67.2 & 63.67 & 58.99 & 61.97 \\
|
| 42 |
+
% -- MaxPool & 59.86 & 74.89 & 85.25 & 66.78 & 61.45 & 59.40 & 60.76 \\
|
| 43 |
+
% -- Coarse & 56.92 & 72.55 & 82.91 & 64.49 & 62.83 & 57.42 & 54.47 \\ \hline
|
| 44 |
+
% \end{tabular}
|
| 45 |
+
% \addtolength{\tabcolsep}{2pt}
|
| 46 |
+
% }
|
| 47 |
+
% \vspace{-3mm}
|
| 48 |
+
% \caption{Ablation study of our model components.
|
| 49 |
+
% % \yq{This table can be made one-column to save space.}\cezhang{+1, we're running out of space.}
|
| 50 |
+
% }
|
| 51 |
+
% \label{table:component_ablation}
|
| 52 |
+
% \vspace{-2mm}
|
| 53 |
+
% \end{table}
|
paper-tex/tables/stats.tex
ADDED
|
@@ -0,0 +1,22 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\begin{table}[h!]
|
| 3 |
+
\centering
|
| 4 |
+
\resizebox{\linewidth}{!}{
|
| 5 |
+
\renewcommand{\arraystretch}{1.1}
|
| 6 |
+
\begin{tabular}{r|cccc}
|
| 7 |
+
\hline
|
| 8 |
+
|
| 9 |
+
Dataset & \begin{tabular}[x]{@{}c@{}}\# Wire\\Images\end{tabular} & \begin{tabular}[x]{@{}c@{}}Min.\\Image Size\end{tabular} &\begin{tabular}[x]{@{}c@{}}Max.\\Image Size\end{tabular} &\begin{tabular}[x]{@{}c@{}}Median\\Image Size\end{tabular} \\ \hline
|
| 10 |
+
Powerline~\cite{powerlinedataset} & 2000 & 128$\times$128 & 128$\times$128 & 128$\times$128\\
|
| 11 |
+
PLDU~\cite{pldu} & 573 & 540$\times$360 & 540$\times$360 & 540$\times$360 \\
|
| 12 |
+
PLDM~\cite{pldu} & 287 & 540$\times$360 & 540$\times$360 & 540$\times$360 \\
|
| 13 |
+
TTPLA~\cite{ttpla} & 1100 & 3840$\times$2160 & 3840$\times$2160 & 3840$\times$2160\\ \hline
|
| 14 |
+
\textbf{Ours} & 6000 & 360$\times$240 & 15904$\times$10608 & 5040$\times$3360 \\ \hline
|
| 15 |
+
|
| 16 |
+
\end{tabular}
|
| 17 |
+
}
|
| 18 |
+
\vspace{-2mm}
|
| 19 |
+
\caption{Statistics of our wire dataset compared to others.}%Image and annotation statistics of our test set.\yq{You can make it as a 4x4 table, with each row: item, min, max, avg; and each column: item name, image size, wire thickness, and percentage. (mt: done)}}
|
| 20 |
+
\vspace{-5mm}
|
| 21 |
+
\label{table:stats}
|
| 22 |
+
\end{table}
|
paper-tex/tables/thresholds.tex
ADDED
|
@@ -0,0 +1,22 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
%auto-ignore
|
| 2 |
+
\begin{table}[t!]
|
| 3 |
+
\centering
|
| 4 |
+
\resizebox{0.9\columnwidth}{!}{
|
| 5 |
+
\renewcommand{\arraystretch}{1.1}
|
| 6 |
+
\begin{tabular}{r|c|ccc|cc}
|
| 7 |
+
\hline
|
| 8 |
+
$\alpha$& \begin{tabular}[x]{@{}c@{}}Wire\\IoU\end{tabular} & F1 & Precision & Recall & \begin{tabular}[x]{@{}c@{}}Avg. Time\\(s/img)\end{tabular} &
|
| 9 |
+
Speed up\\ \hline
|
| 10 |
+
0.0 & 60.97 & 75.75 & 82.63 & 69.93 & 1.91 & 1$\times$ \\
|
| 11 |
+
0.01 & 60.83 & 75.65 & 83.62 & 69.06 & 0.82 & 2.3$\times$ \\
|
| 12 |
+
0.02 & 60.35 & 75.27 & 83.97 & 68.20 & 0.75 & 2.5$\times$ \\
|
| 13 |
+
0.05 & 55.17 & 71.11 & 84.84 & 61.20 & 0.58 & 3.3$\times$ \\
|
| 14 |
+
0.1 & 42.44 & 59.59 & 86.06 & 45.57 & 0.4 & 4.8$\times$ \\
|
| 15 |
+
\hline
|
| 16 |
+
\end{tabular}
|
| 17 |
+
}
|
| 18 |
+
\vspace{-1mm}
|
| 19 |
+
\caption{Ablation on the threshold for refinement. At $\alpha=0.0$, all windows are passed to the fine module.}
|
| 20 |
+
\label{table:thresholds}
|
| 21 |
+
%\vspace{-2mm}
|
| 22 |
+
\end{table}
|