Learning Text Representations for Finding Similar Exercises

Mathematical Intelligent Tutor System brings great convenience for both teachers and students. A basic task in the system is to find similar exercises, which examine students the same skills or knowledge. Inspired by previous work, we propose a new model called Siamese based Bidirectional Encoder Representations from Transformer (SBERT). After training on our Chinese math exercises dataset, AUC(Area Under Curve) of SBERT model can reach up to 0.90, which is higher than that of existed models. Visualization analysis also proves that our model obtains better text representing performance of exercises than previous work.