TRELM: Towards Robust and Efficient Pre-training for Knowledge-Enhanced Language Models