This paper proposes a method called simulated+unsupervised (s+u) learning to enhance the realism of synthetic images by using unlabeled real data, addressing the gap in performance between synthetic and real image distributions. The authors introduce a refiner network that utilizes an adversarial loss and self-regularization to generate realistic images while preserving existing annotation information, ultimately yielding state-of-the-art results in tasks such as gaze and hand pose estimation. Key innovations include local adversarial losses and updates to the discriminator using a history of refined images to stabilize training and mitigate artifacts.