Virtual reality (VR) is a technology that relies on a computer graphics system and other external display and control interfaces, to create an immersive experience by generating an interactive three-dimensional environment on a computer. Currently, however, most virtual reality scenes are far behind the real world in naturalism. One of the limitations is the insufficient graphics computing performance of the computer. It is difficult for mainstream consumer GPUs to meet the requirements of high picture quality and high fluency at the same time when running VR scenes, resulting in a reduction in the game’s visual experience and even human discomfort. In order to balance the quality and fluency of the picture, the areas within and outside the focus range of the user’s sight can be rendered hierarchically, so as to efficiently use computing resources. In order to achieve this goal, the following article proposes a model that combines the saliency information of the virtual scene and the head motion information to predict the focus of the field of view in real time. The model can assign different rendering priorities to objects in the field of view according to the prediction results, and give different priorities, use different rendering algorithms to provide a flexible VR scene rendering optimization solution.