2D virtual try-on has become a hot topic in recent years. It can change what a person image wearing by inputting a desired clothes image. In this study, we propose a visual try-on network, namely Semantic-guided and Detail-preserving Image-based 2D Virtual Try-On Networks(SD-VTON), which improve the architecture of a novel network, ACGPN. Considering the three major modules of ACGPN, including the Semantic Generation Module(SGM), the Clothes Warping Module (CWM), and the Content Fusion Module(CFM), we find out that SGM is the main reason causing the problems mentioned above. Consequently, we substitute UNet++ for the original neural network to improve human parsing, and hoping it makes the same or even better result with fewer epochs. In the test result of 200 epochs, there are 86.1% SSIM scores higher than ACGPN’s. Moreover, the result of 140 epochs is extremely close to that of 200 epochs, which means that we can save more time on training. In terms of practicality, SD-VTON add a Clothes Tailoring Module at the head of the overall architecture, generating semantic segmentation with JPPNet, and develop a virtual try-on system with higher quality.