cosFormer: Rethinking Softmax in Attention