dc.contributor.author |
Gunasekara, CM |
|
dc.date.accessioned |
2024-07-18T09:05:23Z |
|
dc.date.available |
2024-07-18T09:05:23Z |
|
dc.date.issued |
2023-12 |
|
dc.identifier.uri |
http://dl.lib.uom.lk/handle/123/22573 |
|
dc.description.abstract |
Object detection and filtering based on shape and color is an important capability for many robotics applications. For example, sorting objects by shape and color is a common industrial application. Service robots also need to detect and track objects based on visual properties. While powerful deep learning approaches like YOLO have emerged for general object detection, they require large datasets and extensive training. A simple shape and color filtering provide a lightweight and customizable alternative.
This work aims to provide a real-time modular and lightweight system that can identify objects of basic shapes and colors and allow extensibility of functionality by incorporating more custom color and shape filters. The proposed system for real-time shape and color filtering using a Microsoft Kinect RGB-D sensor and Robot Operating System (ROS2) can identify an array of regular shapes like circles, rectangles, and triangles over a spectrum of different colors. New shape and color filters can be added dynamically at runtime thanks to the modular, ROS-2-based implementation. |
en_US |
dc.language.iso |
en |
en_US |
dc.publisher |
Engineering Research Unit |
en_US |
dc.subject |
Kinect |
en_US |
dc.subject |
ROS2 |
en_US |
dc.subject |
Real-time |
en_US |
dc.subject |
Shape filter |
en_US |
dc.subject |
Color filter |
en_US |
dc.title |
A Real-time, scalable and extensible object filtering and detection system using kinect sensor and ROS2 foxy |
en_US |
dc.type |
Conference-Extended-Abstract |
en_US |
dc.identifier.faculty |
Engineering |
en_US |
dc.identifier.department |
Department of Computer Science and Engineering |
en_US |
dc.identifier.year |
2023 |
en_US |
dc.identifier.conference |
ERU Symposium - 2023 |
en_US |
dc.identifier.place |
Sri Lanka |
en_US |
dc.identifier.pgnos |
pp. 40-41 |
en_US |
dc.identifier.proceeding |
Proceedings of the ERU Symposium 2023 |
en_US |
dc.identifier.doi |
https://doi.org/10.31705/ERU.2023.19 |
en_US |