center  on  the  panoramic  image  is  derived.  It  was 
shown  analytically  that  the  inclination  may  be 
approximated by a sine wave. 
In A, the sine wave is fitted by the least squares 
method, whereas in B, robust and accurate fitting is 
realized by eliminating outliers using RANSAC. 
In A, the estimation accuracy is 3.3° for the first 
correction and 1.5° for the second one, but in B, an 
estimation accuracy of 0.1° is obtained without the 
need for a second step (when the effect of noise can 
be eliminated). 
For A, it was reported that there were no failure 
examples in an experiment of 40 examples, but only 
one example  was  shown. For  B,  many experiments 
were  carried  out  in  various  environments,  and 
concrete examples are shown. 
7  CONCLUSIONS 
In  this  paper,  we  proposed  a  method  for  upright 
adjustment  of  a  panoramic  image  by  detecting  the 
inclination  of  the  camera  from  a  pre-corrected 
panoramic  image  with  high  accuracy  and  at  high 
speed  using  a  vertical  line  existing  in  an  indoor 
environment  or  an  outdoor  environment  near  a 
building. 
Because of the nature of this method, it cannot be 
used in an environment without vertical lines, but it is 
useful  in  many environments in  which autonomous 
robots are expected to operate in the future, such as 
normal indoor environments, construction sites, and 
in and around warehouses. When the lengths of the 
projection curves are extremely short, the method is 
easily affected by noise, and the upright adjustment 
tends to be unstable. However, even in this case, the 
value of pthresh can be added as a certainty and the 
correct  handling  can  be  performed  in  the  post-
processing. 
In future, this method will be applied to tasks such 
as  the  self-localization  of  autonomous  robots, 
reconstruction  of  3D  environments,  and  object 
recognition.  In  addition,  by  integrating  self-
localization estimation and a 3D environment model, 
speed-up and robustness will be achieved by learning 
the parameters of the  upright adjustment depending 
on location. 
REFERENCES 
Bazin,  J.-C.,  Demonceaux,  C.,  et  al.  2012.  Rotation 
estimation  and  vanishing  point  extraction  by 
omnidirectional  vision  in  urban  environment.  The 
International Journal of Robotics Research 31(1): 63-
81. 
Bosse, M., Rikoski, R., et al. 2002. Vanishing points and 3d 
lines  from  omnidirectional  video.  Proceedings. 
International Conference on Image Processing, IEEE. 
Canny,  John  1986.  A  computational  approach  to  edge 
detection. IEEE Transactions on Pattern Analysis and 
Machine Intelligence, 8(6): 679-698. 
Cipolla,  R.,  Robertson,  D.,  et  al.  1999.  Photobuilder-3D 
models  of  architectural  scenes  from  uncalibrated 
images.  Proceedings IEEE International Conference 
on Multimedia Computing and Systems, IEEE. 
Coughlan, J. M. and Yuille, A. L. 1999. Manhattan world: 
Compass  direction  from  a  single  image  by  bayesian 
inference.  Proceedings of the seventh IEEE 
international conference on computer vision, IEEE. 
Demonceaux, C., Vasseur, P., et al. 2006. Omnidirectional 
vision on UAV  for attitude computation. Proceedings 
2006 IEEE International Conference on Robotics and 
Automation, 2006. ICRA 2006., IEEE. 
Demonceaux,  C.,  Vasseur,  P.,  et  al.  2007.  UAV  attitude 
computation  by  omnidirectional  vision  in  urban 
environment.  Proceedings 2007 IEEE International 
Conference on Robotics and Automation, IEEE. 
Fischler, A., Martin and Bolles, C., Robert 1981. Random 
sample  consensus:  a  paradigm  for  model  fitting  with 
applications  to  image  analysis  and  automated 
cartography. Communications of the ACM, 24(6): 381-
395. 
Gallagher,  A.  C.  2005.  Using  vanishing  points  to  correct 
camera  rotation  in  images.  The 2nd Canadian 
Conference on Computer and Robot Vision (CRV'05), 
IEEE. 
Illingworth, John and Kittler, Josef 1988. A survey of the 
Hough  transform.  Computer vision, graphics and 
image processing, 44(1): 87-116. 
Jayasuriya,  M.,  Ranasinghe,  R.,  et  al.  2020.  Active 
Perception  for  Outdoor  Localisation  with  an 
Omnidirectional  Camera.  IEEE/RSJ International 
Conference on Intelligent Robots and Systems. 
Jeon, J., Jung, J., et al. 2018. Deep Upright Adjustment of 
360 Panoramas Using Multiple Roll Estimations. Asian 
Conference on Computer Vision, Springer. 
Joo, K., Oh, T.-H., et al. 2018. Globally optimal inlier set 
maximization  for  Atlanta  frame  estimation. 
Proceedings of the IEEE Conference on Computer 
Vision and Pattern Recognition. 
Jung, J., Kim, B., et al. 2017. Robust upright adjustment of 
360 spherical  panoramas.  The Visual Computer 33(6-
8): 737-747. 
Jung,  R.,  Lee,  A.  S.  J.,  et  al. 2019.  Deep360Up:  A Deep 
Learning-Based  Approach  for  Automatic  VR  Image 
Upright Adjustment. 2019 IEEE Conference on Virtual 
Reality and 3D User Interfaces (VR), IEEE. 
Kawai,  N.  2019.  A  method  for  rectifying  inclination  of 
panoramic  images.  ACM SIGGRAPH 2019 Posters, 
ACM. 
Lee,  H.,  Shechtman,  E.,  et  al.  2013.  Automatic  upright 
adjustment  of  photographs  with  robust  camera