Planned maintenance
A system upgrade is planned for 10/12-2024, at 12:00-13:00. During this time DiVA will be unavailable.
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Deep Direct Visual Odometry
Key Laboratory of Smart Manufacturing in Energy Chemical Process, Ministry of Education, East China University of Science and Technology, Shanghai 200237, China.
Key Laboratory of Smart Manufacturing in Energy Chemical Process, Ministry of Education, East China University of Science and Technology, Shanghai 200237, China.
Key Laboratory of Smart Manufacturing in Energy Chemical Process, Ministry of Education, East China University of Science and Technology, Shanghai 200237, China.
Luleå University of Technology, Department of Computer Science, Electrical and Space Engineering, Computer Science. School of Electrical and Data Engineering, University of Technology Sydney, Sydney, NSW, Australia; College of Mathematics and Computer Science, Fuzhou University, Fuzhou 350116, China.ORCID iD: 0000-0003-1902-9877
2022 (English)In: IEEE transactions on intelligent transportation systems (Print), ISSN 1524-9050, E-ISSN 1558-0016, Vol. 23, no 7, p. 7733-7742Article in journal (Refereed) Published
Abstract [en]

Traditional monocular direct visual odometry (DVO) is one of the most famous methods to estimate the ego-motion of robots and map environments from images simultaneously. However, DVO heavily relies on high-quality images and accurate initial pose estimation during tracking. With the outstanding performance of deep learning, previous works have shown that deep neural networks can effectively learn 6-DoF (Degree of Freedom) poses between frames from monocular image sequences in the unsupervised manner. However, these unsupervised deep learning-based frameworks cannot accurately generate the full trajectory of a long monocular video because of the scale-inconsistency between each pose. To address this problem, we use several geometric constraints to improve the scale-consistency of the pose network, including improving the previous loss function and proposing a novel scale-to-trajectory constraint for unsupervised training. We call the pose network trained by the proposed novel constraint as TrajNet. In addition, a new DVO architecture, called deep direct sparse odometry (DDSO), is proposed to overcome the drawbacks of the previous direct sparse odometry (DSO) framework by embedding TrajNet. Extensive experiments on the KITTI dataset show that the proposed constraints can effectively improve the scale-consistency of TrajNet when compared with previous unsupervised monocular methods, and integration with TrajNet makes the initialization and tracking of DSO more robust and accurate.

Place, publisher, year, edition, pages
IEEE, 2022. Vol. 23, no 7, p. 7733-7742
Keywords [en]
Visual odometry, direct methods, pose estimation, deep learning, unsupervised learning
National Category
Computer Vision and Robotics (Autonomous Systems)
Research subject
Pervasive Mobile Computing
Identifiers
URN: urn:nbn:se:ltu:diva-84133DOI: 10.1109/TITS.2021.3071886ISI: 000732877700001Scopus ID: 2-s2.0-85104622568OAI: oai:DiVA.org:ltu-84133DiVA, id: diva2:1549265
Note

Validerad;2022;Nivå 2;2022-08-19 (sofila);

Finansiär:  National Natural Science Foundation of China (61988101); International (Regional) Cooperation and Exchange Project (61720106008); Program of Shanghai Academic Research Leader (20XD1401300); Programme of Introducing Talents of Discipline to Universities (B17017)

Available from: 2021-05-05 Created: 2021-05-05 Last updated: 2022-08-19Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Vasilakos, Athanasios V.

Search in DiVA

By author/editor
Vasilakos, Athanasios V.
By organisation
Computer Science
In the same journal
IEEE transactions on intelligent transportation systems (Print)
Computer Vision and Robotics (Autonomous Systems)

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 58 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf