Video semantic segmentation with low latency
Dublin Core
Title
Video semantic segmentation with low latency
Subject
Convolutional neural network
Decision network
FlowNet
Latency
Object detection
SegNet
Semantic segmentation
Decision network
FlowNet
Latency
Object detection
SegNet
Semantic segmentation
Description
Recent advances in computer vision and deep learning algorithms have yielded intriguing results. It can perform tasks previously requiring human eyes and brains. Semantic video segmentation for autonomous cars is difficult due to the high cost, low latency, and performance requirements of convolutional neural networks (CNNs). Deep learning architectures like SegNet and FlowNet 2.0 on the Cambridge-driving labeled video database (CamVid) dataset enable low-latency pixel-wise semantic segmentation of video features. Because it uses SegNet and FlowNet topologies, it is ideal for practical applications. The decision network chooses an optical flow or segmentation network for an image frame based on the expected confidence score. Combining this decision-making method with adaptive scheduling of the key frame approach can speed up the process. ResNet50 SegNet has a “54.27%” mean intersection over union (MIoU) and a “19.57” average FPS. In addition to decision network and adaptive key frame sequencing, FlowNet2.0 increased graphics processing unit (GPU) frame processing per second to “30.19” with a MIoU of “47.65%”. The GPU is used “47.65%” of the time. This performance gain illustrates that the video semantic segmentation network is faster without sacrificing quality.
Creator
Channappa Gowda D. V., Kanagavalli R.
Source
Journal homepage: http://telkomnika.uad.ac.id
Date
Mar 26, 2024
Contributor
PERI IRAWAN
Format
PDF
Language
ENGLISH
Type
TEXT
Files
Collection
Citation
Channappa Gowda D. V., Kanagavalli R., “Video semantic segmentation with low latency,” Repository Horizon University Indonesia, accessed January 11, 2026, https://repository.horizon.ac.id/items/show/10262.