In response to the issue of reduced road environment perception accuracy for intelligent vehicles under extreme haze conditions, this paper proposed a joint haze environment perception algorithm based on an improved CycleGAN and YOLOv8. Firstly, the CycleGAN algorithm was used as the framework for image defogging preprocessing. A self-attention mechanism was incorporated into the generator network to enhance the network’s feature extraction capability. Additionally, to minimize color discrepancies with real images, a self-regularized color loss function was introduced. Secondly, in the object detection phase, the lightweight GhostConv network was first used to replace the original backbone network, reducing computational complexity. Furthermore, the GAM attention mechanism was added to the neck network to effectively improve the network’s ability to interact with global information. Finally, the WIoU loss function was used to mitigate harmful gradients caused by low-quality samples, improving the model’s convergence speed. Experiments conducted on the RESIDE and BDD100k datasets validate the proposed algorithm. Results show that the structural similarity between dehazed and original images is 85%. Compared to the original CycleGAN algorithm and the AODNet algorithm, the proposed approach improves the peak signal-to-noise ratio (PSNR) by 2.24 dB and 2.5 dB, respectively, and the structural similarity index (SSIM) by 15.4% and 36.3%, respectively. Additionally, the improved YOLOv8 algorithm demonstrates enhancements over the original algorithm, with precision, recall, and mean average precision (mAP) increasing by 2.5%, 1.8%, and 1.1%, respectively. The experimental results confirm that the proposed algorithm outperforms traditional algorithms in terms of recall and detection accuracy, demonstrating its practical value