We propose VLRM, a method for fine-tuning an existing image captioning model using reinforcement learning and vision-language models as reward models. The method manages to significantly improve the generation quality without human-labeled data and is applicable to any image captioning model. Our model reaches impressive 0.90 R@1 CLIP Recall score on MS-COCO Carpathy Test Split.