The paper contributes to No-Reference video quality assessment of broadcasted HD video over IP networks and DVB. In this work we have enhanced our bottom-up spatio-temporal saliency map model by considering semantics of the visual scene. Thus we propose a new saliency map model based on face detection that we called semantic saliency map. A new fusion method has been proposed to merge the bottom-up saliency maps with the semantic saliency map. We show that our NR metric WMBER weighted by the spatio-temporal-semantic saliency map provides higher results then the WMBER weighted by the bottom-up spatio-temporal saliency map. Tests are performed on two H.264/AVC video databases for video quality assessment over lossy networks.