Autonomous platforms such as self-driving vehicles, advanced driver-assistance systems (ADAS), and intelligent aerial drones demand real-time video perception systems capable of delivering actionable visual information at ultra-low latency. High-resolution vision pipelines are often hindered by delays introduced at multiple stages-sensor acquisition, video encoding, data transmission, decoding, and display-undermining the responsiveness required for safety-critical decision making. This study introduces a holistic system-level optimization framework that systematically reduces end-to-end video latency while maintaining image fidelity and perception accuracy. The proposed approach integrates hardware-accelerated encoding, zero-copy direct memory access (DMA), lightweight UDP-based RTP transport, and GPU-accelerated decoding into a unified pipeline. By minimizing redundant memory copies and software bottlenecks, the system achieves seamless data flow across hardware and software boundaries. Evaluations demonstrate a latency reduction from a baseline of 45.3 milliseconds to an optimized 23.5 milliseconds, representing a 48.1% improvement without sacrificing spatial resolution or detection robustness. Under optimized configurations, the framework sustains frame rates above 60 FPS at both Full HD and 4K resolutions, with frame drop rates held to approximately 3%. Perceptual evaluation further confirms that object detection accuracy consistently exceeds 91% within the <35 ms latency range, while collision-prediction delays are reduced to below 12.4 ms, ensuring timely responses in dynamic scenarios. These improvements collectively validate the critical importance of hardware-software co-design for embedded vision systems. The results highlight that ultra-low-latency perception is achievable on edge platforms when pipelines are designed with cross-layer optimization, bridging sensor interfaces, video codecs, network transport, and GPU computation. The proposed architecture provides a scalable foundation for future embedded vision deployments in autonomous driving, robotics, and unmanned aerial systems, where low latency is a non-negotiable requirement for safety, reliability, and operational efficiency.