Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Towards Deep-learning-based Autonomous Navigation in the Short-loading Cycle
Luleå University of Technology, Department of Computer Science, Electrical and Space Engineering, Embedded Internet Systems Lab.ORCID iD: 0000-0002-4716-9765
2023 (English)Doctoral thesis, comprehensive summary (Other academic)
Place, publisher, year, edition, pages
Luleå: Luleå University of Technology, 2023.
Series
Doctoral thesis / Luleå University of Technology 1 jan 1997 → …, ISSN 1402-1544
National Category
Computer Sciences Robotics
Research subject
Cyber-Physical Systems
Identifiers
URN: urn:nbn:se:ltu:diva-102486ISBN: 978-91-8048-442-8 (print)ISBN: 978-91-8048-443-5 (electronic)OAI: oai:DiVA.org:ltu-102486DiVA, id: diva2:1812757
Public defence
, Luleå tekniska universitet, Luleå (English)
Supervisors
Available from: 2023-11-17 Created: 2023-11-17 Last updated: 2023-11-17Bibliographically approved
List of papers
1. Machine Vision for Construction Equipment by Transfer Learning with Scale Models
Open this publication in new window or tab >>Machine Vision for Construction Equipment by Transfer Learning with Scale Models
2020 (English)In: 2020 International Joint Conference on Neural Networks (IJCNN), IEEE, 2020, article id 21108Conference paper, Published paper (Refereed)
Abstract [en]

Machine vision is required by autonomous heavy construction equipment to navigate and interact with the environment. Wheel loaders need the ability to identify different objects and other equipment to perform the task of automatically loading and dumping material on dump trucks, which can be achieved using deep neural networks. Training such networks from scratch requires the iterative collection of potentially large amounts of video data, which is challenging at construction sites because of the complexity of safely operating heavy equipment in realistic environments. Transfer learning, for which pretrained neural networks can be retrained for use at construction sites, is thus attractive, especially if data can be acquired without full-scale experiments. We investigate the possibility of using scalemodel data for training and validating two different pretrained networks and use real-world test data to examine their generalization capability. A dataset containing 268 images of a 1:16 scale model of a Volvo A60H dump truck is provided, as well as 64 test images of a full-size Volvo A25G dump truck. The code and dataset are publicly available 1 . The networks, both pretrained on the MS-COCO dataset, were fine-tuned to the created dataset, and the results indicate that both networks can learn the features of the scale-model dump truck (validation mAP of 0.82 for YOLOv3 and 0.95 for RetinaNet). Both networks can transfer these learned features to detect objects on a full-size dump truck with no additional training (test mAP of 0.70 for YOLOv3 and 0.79 for RetinaNet).

Place, publisher, year, edition, pages
IEEE, 2020
Series
International Joint Conference on Neural Networks (IJCNN), ISSN 2161-4393, E-ISSN 2161-4407
Keywords
construction equipment, automation, computer vision, deep learning, machine learning
National Category
Embedded Systems
Research subject
Cyber-Physical Systems
Identifiers
urn:nbn:se:ltu:diva-81008 (URN)10.1109/IJCNN48605.2020.9207577 (DOI)000626021407089 ()2-s2.0-85093872275 (Scopus ID)
Conference
2020 International Joint Conference on Neural Networks (IJCNN), 19-24 July, 2020, Glasgow, United Kingdom
Note

ISBN för värdpublikation: 978-1-7281-6926-2, 978-1-7281-6927-9

Available from: 2020-10-02 Created: 2020-10-02 Last updated: 2023-11-17Bibliographically approved
2. Deep-learning-based vision for earth-moving automation
Open this publication in new window or tab >>Deep-learning-based vision for earth-moving automation
2022 (English)In: Automation in Construction, ISSN 0926-5805, E-ISSN 1872-7891, Vol. 133, article id 104013Article, review/survey (Refereed) Published
Abstract [en]

Earth-moving machines are heavy-duty vehicles designed for construction operations involving earthworks. The tasks performed by such machines typically involve navigation and interaction with materials such as soil, gravel, and blasted rock. Skilled operators use a combination of visual, sound, tactile and possibly motion feedback to perform tasks efficiently. We survey the literature in this research area and analyse the relative importance of different sensor system modalities focusing on deep-learning-based vision and automation for the short-cycle loading task. This is a common and repetitive task that is attractive to automate. The analysis indicates that computer vision, in combination with onboard sensors, is more critical than coordinate-based positioning. Furthermore, we find that data-driven approaches, in general, have high potential in terms of productivity, adaptability, versatility and wear and tear with respect to automation system solutions. The main knowledge gaps identified relate to loading non-fine heterogeneous material and navigation during loading and unloading.

Place, publisher, year, edition, pages
Elsevier, 2022
Keywords
Automation, Short-cycle Loading, Construction equipment, Computer vision, Machine learning
National Category
Computer Sciences
Research subject
Cyber-Physical Systems; Machine Learning
Identifiers
urn:nbn:se:ltu:diva-87710 (URN)10.1016/j.autcon.2021.104013 (DOI)000716642800006 ()2-s2.0-85118161462 (Scopus ID)
Note

Validerad;2021;Nivå 2;2021-11-01 (johcin)

Available from: 2021-11-01 Created: 2021-11-01 Last updated: 2023-11-17Bibliographically approved
3. Semi-Automatic Video Frame Annotation for Construction Equipment Automation Using Scale-Models
Open this publication in new window or tab >>Semi-Automatic Video Frame Annotation for Construction Equipment Automation Using Scale-Models
2021 (English)In: IECON 2021 – 47th Annual Conference of the IEEE Industrial Electronics Society, IEEE, 2021Conference paper, Published paper (Refereed)
Abstract [en]

Data collection and annotation is a time consuming and costly process, yet necessary for machine vision. Automation of construction equipment relies on seeing and detecting different objects in the vehicle’s surroundings. Construction equipment is commonly used to perform frequent repetitive tasks, which are interesting to automate. An example of such a task is the short-loading cycle, where the material is moved from a pile into the tipping body of a dump truck for transport. To complete this task, the wheel loader needs to have the capability to locate the tipping body of the dump truck. The machine vision system also allows the vehicle to detect unforeseen dangers such as other vehicles and more importantly human workers. In this work, we investigate the viability to perform semi-automatic annotation of video data using linear interpolation. The data is collected using scale-models mimicking a wheel-loaders approach towards a dump truck during the short-loading cycle. To measure the viability of this type of solution, the workload is compared to the accuracy of the model, YOLOv3. The results indicate that it is possible to maintain the performance while decreasing the annotation workload by about 95%. This is an interesting result for this application domain, as safety is critical and retaining the vision system performance is more important than decreasing the annotation workload. The fact that the performance seems to retain with a large workload decrease is an encouraging sign.

Place, publisher, year, edition, pages
IEEE, 2021
Keywords
Autonomous construction equipment, semi-automatic annotation, video-stream data, object detection, linear interpolation
National Category
Computer Vision and Robotics (Autonomous Systems)
Research subject
Cyber-Physical Systems; Machine Learning
Identifiers
urn:nbn:se:ltu:diva-87866 (URN)10.1109/iecon48115.2021.9589255 (DOI)000767230601030 ()2-s2.0-85119470296 (Scopus ID)
Conference
47th Annual Conference of the IEEE Industrial Electronics Society (IECON 2021), 13-16 Oct. 2021,Toronto, ON, Canada
Note

ISBN för värdpublikation:978-1-6654-3554-3, 978-1-6654-0256-9

Available from: 2021-11-11 Created: 2021-11-11 Last updated: 2023-11-17Bibliographically approved
4. Autonomous Navigation of Wheel Loaders using Task Decomposition and Reinforcement Learning
Open this publication in new window or tab >>Autonomous Navigation of Wheel Loaders using Task Decomposition and Reinforcement Learning
2023 (English)Conference paper, Published paper (Refereed)
Abstract [en]

The short-loading cycle is a repetitive task performed in high quantities making it a good candidate for automation. Expert operators perform this task to upkeep high productivity while minimizing the environmental impact of using energy to propel the wheel loader. The need to balance productivity and environmental performance is essential for the sub-task of navigating the wheel loader between the pile of material and a dump truck receiving the material. This task is further complicated by behaviours of the wheel loader such as wheel slip depending on the tire-to-surface friction that is hard to model. Such uncertainties motivate the use of data-driven and adaptable approaches like reinforcement learning to automate navigation. In this paper, we examine the possibility to use reinforcement learning for the navigation sub-task. We focus on the process of developing a solution to the complete sub-task by decomposing it into two distinct steps and training two different agents to perform them separately. These steps are reversing from the pile and approaching the dump truck. The agents are trained in a simulation environment in which the wheel loader is modelled. Our results indicate that task decomposition can be helpful in performing the navigation compared to training a single agent for the entire sub-task. We present unsuccessful experiments using a single agent for the entire sub-task to illustrate difficulties associated with such an approach. A video of the results is available online. Video available at https://youtu.be/IZbgvHvSltI.

National Category
Robotics
Identifiers
urn:nbn:se:ltu:diva-101844 (URN)10.1109/CASE56687.2023.10260481 (DOI)
Conference
2023 IEEE 19th International Conference on Automation Science and Engineering (CASE), Auckland, New Zealand, 2023
Available from: 2023-10-30 Created: 2023-10-30 Last updated: 2023-11-18
5. Automating the Short-Loading Cycle: Challenges, Survey and Integration Framework
Open this publication in new window or tab >>Automating the Short-Loading Cycle: Challenges, Survey and Integration Framework
(English)Manuscript (preprint) (Other academic)
National Category
Robotics Computer Sciences
Identifiers
urn:nbn:se:ltu:diva-101849 (URN)
Available from: 2023-10-30 Created: 2023-10-30 Last updated: 2023-11-17
6. Learning the Approach During the Short-loading Cycle Using Reinforcement Learning
Open this publication in new window or tab >>Learning the Approach During the Short-loading Cycle Using Reinforcement Learning
(English)Manuscript (preprint) (Other academic)
National Category
Robotics Computer Sciences
Identifiers
urn:nbn:se:ltu:diva-101851 (URN)
Available from: 2023-10-30 Created: 2023-10-30 Last updated: 2023-11-17

Open Access in DiVA

No full text in DiVA

Authority records

Borngrund, Carl

Search in DiVA

By author/editor
Borngrund, Carl
By organisation
Embedded Internet Systems Lab
Computer SciencesRobotics

Search outside of DiVA

GoogleGoogle Scholar

isbn
urn-nbn

Altmetric score

isbn
urn-nbn
Total: 16 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf