International Conference on Computer Vision and Graphics ICCVG 2024
Tuesday, 17 September 2024 -
10:00
Monday, 16 September 2024
Tuesday, 17 September 2024
10:00
Opening of the Conference
Opening of the Conference
(main track)
10:00 - 10:05
Room: Online meeting - MS Teams
10:05
Method for Fine Registration of Point Sets Based on the Curvature of the Surface
-
M. Jiřina
(
Department of Applied Mathematics, Czech Technical University in Prague, Prague, Czech Republic
)
J. Glaser
(
Department of Applied Mathematics, Czech Technical University in Prague, Prague, Czech Republic
)
Method for Fine Registration of Point Sets Based on the Curvature of the Surface
(main track)
M. Jiřina
(
Department of Applied Mathematics, Czech Technical University in Prague, Prague, Czech Republic
)
J. Glaser
(
Department of Applied Mathematics, Czech Technical University in Prague, Prague, Czech Republic
)
10:05 - 10:30
Room: Online meeting - MS Teams
Efficient and accurate point set registration is an important task in 3D scene reconstruction in computer vision. This paper presents a method called Curvature Surface Iterative Closest Point (CS-ICP) for precise point set registration. By leveraging the curvature characteristics of the point set input, CS-ICP resolves local minima challenges encountered by standard ICP algorithms, demonstrating superior precision across various datasets. Additionally, CS-ICP significantly reduces computation time by working with fewer points per iteration, reducing the runtime by around 83\% compared to the reference methods. This paper also introduces evaluation criteria based on Euclidean and Chebyshev measures, offering a better assessment of point set registration quality without needing additional attributes such as evaluation ICP threshold.
10:30
Contextual Information-Based Registration Method for Point Set Registration
-
J. Glaser
(
Department of Applied Mathematics, Czech Technical University in Prague, Prague, Czech Republic
)
T. Laurin
(
Department of Applied Mathematics, Czech Technical University in Prague, Prague, Czech Republic
)
M. Jiřina
(
Department of Applied Mathematics, Czech Technical University in Prague, Prague, Czech Republic
)
Contextual Information-Based Registration Method for Point Set Registration
(main track)
J. Glaser
(
Department of Applied Mathematics, Czech Technical University in Prague, Prague, Czech Republic
)
T. Laurin
(
Department of Applied Mathematics, Czech Technical University in Prague, Prague, Czech Republic
)
M. Jiřina
(
Department of Applied Mathematics, Czech Technical University in Prague, Prague, Czech Republic
)
10:30 - 10:55
Room: Online meeting - MS Teams
This paper introduces a method called Contextual Information-Based Registration (CIBR), used to accurately register large and dense point sets, which represent a 3D scene. Distinguished from existing techniques, CIBR works with input point sets by partitioning them into discrete logical parts, which represent objects in the scene. A registration process takes place on each part of the point set, which contains the richest contextual information related to the 3D objects in the point clouds, leading to a final precise alignment. Through experimentation, CIBR demonstrates superior precision across various datasets, with the best improvement of 267% in fitness and correspondence set size and 52.2% in inliner RMSE, even though there exist cases, when the registration was suboptimal. CIBR achieves in most cases more robust and precise registration outcomes than the traditional fine and rough ICP registration methods.
10:55
An example of the use of a small dataset for the classification of simple actions based on manually extracted shape descriptors, a single-layer neural network and leave-one-actor-out cross-validation procedure
-
Dariusz Frejlichowski
(
West Pomeranian University of Technology, Szczecin, Poland
)
Katarzyna Gościewska
(
West Pomeranian University of Technology, Szczecin, Poland
)
An example of the use of a small dataset for the classification of simple actions based on manually extracted shape descriptors, a single-layer neural network and leave-one-actor-out cross-validation procedure
(main track)
Dariusz Frejlichowski
(
West Pomeranian University of Technology, Szczecin, Poland
)
Katarzyna Gościewska
(
West Pomeranian University of Technology, Szczecin, Poland
)
10:55 - 11:20
Room: Online meeting - MS Teams
In this study, the method that recognizes human actions is analysed. It uses manually created shape features, and is tested in combination with a neural network-based classifier. We assume an application scenario involving the recognition of physical exercises as one of the preventive measures to reduce the risk of non-communicable diseases in the elderly. A popular action recognition dataset is used, as it contains activities corresponding to selected exercises. In addition to the application, the paper focuses on the study of combining neural network classifier with manually created shape features extracted from a small dataset. The main steps of the approach include calculating shape descriptors for all moving foreground objects extracted from video frames, then using these descriptors to construct feature vectors, and ultimately applying the Fourier transform to create representations of action sequences. A coarse classification step is included, which distinguishes between actions performed in place and actions with changing object's locations. The final classification is carried out using a neural network and a leave-one-actor-out cross-validation procedure. This paper presents experimental results on classification using the proposed approach based on simple shape descriptors and a feed-forward neural network with a single hidden layer. The averaged accuracy exceeds 97%.
11:20
Break
Break
(main track)
11:20 - 11:25
Room: Online meeting - MS Teams
11:25
Improving the efficiency of "Show and Tell" encoder-decoder image captioning model
-
Albert Ziółkiewicz
Karol Zieliński
Marcin Iwanowski
(
Institute of Control and Industrial Electronics, Warsaw University of Technology
)
Mateusz Bartosiewicz
(
Institute of Control and Industrial Electronics, Warsaw University of Technology
)
Piotr Szczepański
Improving the efficiency of "Show and Tell" encoder-decoder image captioning model
(main track)
Albert Ziółkiewicz
Karol Zieliński
Marcin Iwanowski
(
Institute of Control and Industrial Electronics, Warsaw University of Technology
)
Mateusz Bartosiewicz
(
Institute of Control and Industrial Electronics, Warsaw University of Technology
)
Piotr Szczepański
11:25 - 11:50
Room: Online meeting - MS Teams
The paper investigates the influence of hyperparameters of the "Show and Tell" image captioning model on the overall efficiency of the method. The method is based on an encoder-decoder approach, where the encoder -- the backbone feature extractor based on the convolutional neural networks (CNN) is responsible for extracting image features and the decoder -- the recurrent neural network (RNN), produces a caption -- a phrase describing the image content. In our research, we tested the encoder part by verifying Densenet, Resnet, and Regnet image feature extractors and the decoder part by changing the size of the RNN sizes. Furthermore, we also investigated the sentence generation stage. The investigation aims to find the optimal feature extractor and decoder size combination. Our research proves that an optimal choice of model's hyperparameters increases caption generation efficiency.
11:50
Utilisation of Vision Systems and Digital Twin for Maintaining Cleanliness in Public Spaces
-
Michał Daniłowicz
(
Embedded Vision Systems Group, Computer Vision Laboratory, Department of Automatic Control and Robotics, AGH University of Krakow, Poland
)
Mateusz Wąsala
(
Embedded Vision Systems Group, Computer Vision Laboratory, Department of Automatic Control and Robotics, AGH University of Krakow, Poland
)
Krzysztof Błachut
(
Embedded Vision Systems Group, Computer Vision Laboratory, Department of Automatic Control and Robotics, AGH University of Krakow, Poland
)
Tomasz Kryjak
(
Embedded Vision Systems Group, Computer Vision Laboratory, Department of Automatic Control and Robotics, AGH University of Krakow, Poland
)
Marcin Kowalczyk
(
Embedded Vision Systems Group, Computer Vision Laboratory, Department of Automatic Control and Robotics, AGH University of Krakow, Poland
)
Hubert Szolc
(
Embedded Vision Systems Group, Computer Vision Laboratory, Department of Automatic Control and Robotics, AGH University of Krakow, Poland
)
Utilisation of Vision Systems and Digital Twin for Maintaining Cleanliness in Public Spaces
(main track)
Michał Daniłowicz
(
Embedded Vision Systems Group, Computer Vision Laboratory, Department of Automatic Control and Robotics, AGH University of Krakow, Poland
)
Mateusz Wąsala
(
Embedded Vision Systems Group, Computer Vision Laboratory, Department of Automatic Control and Robotics, AGH University of Krakow, Poland
)
Krzysztof Błachut
(
Embedded Vision Systems Group, Computer Vision Laboratory, Department of Automatic Control and Robotics, AGH University of Krakow, Poland
)
Tomasz Kryjak
(
Embedded Vision Systems Group, Computer Vision Laboratory, Department of Automatic Control and Robotics, AGH University of Krakow, Poland
)
Marcin Kowalczyk
(
Embedded Vision Systems Group, Computer Vision Laboratory, Department of Automatic Control and Robotics, AGH University of Krakow, Poland
)
Hubert Szolc
(
Embedded Vision Systems Group, Computer Vision Laboratory, Department of Automatic Control and Robotics, AGH University of Krakow, Poland
)
11:50 - 12:15
Room: Online meeting - MS Teams
Nowadays, the increasing demand for maintaining high cleanliness standards in public spaces results in the search for innovative solutions. The deployment of CCTV systems equipped with modern cameras and software enables not only real-time monitoring of the cleanliness status but also automatic detection of impurities and optimisation of cleaning schedules. The Digital Twin technology allows for the creation of a virtual model of the space, facilitating the simulation, training, and testing of cleanliness management strategies before implementation in the real world. In this paper, we present the utilisation of advanced vision surveillance systems and the Digital Twin technology in cleanliness management, using a railway station as an example. The Digital Twin was created based on an actual 3D model in the Nvidia Omniverse Isaac Sim simulator. A litter detector, bin occupancy level detector, stain segmentation, and a human detector (including the cleaning crew) along with their movement analysis were implemented. A preliminary assessment was conducted, and potential modifications for further enhancement and future development of the system were identified.
12:15
Closing of the Conference
Closing of the Conference
(main track)
12:15 - 12:20
Room: Online meeting - MS Teams