Capsule Based Neural Network Architecture to perform completeness check for Patent Eligibility Process

In the process of filing patents, attorneys need to ask many questions, to the inventors, to ascertain patent eligibility. We propose to ease up such conversation through a deep learning-based system. This system can automatically check whether all key ingredients required for checking the patent eligibility are present in technical write-up shared by the inventors. If not, the inventors can provide the missing information. We present a trainable model to identify various ingredients such as the objective, motivation, new observation, etc. from research articles. We model this as a sentence classification problem, which is a difficult task because a patent can be filed in any domain, and sentences involved can often be very long. To this end, we propose a dilated LSTM and capsule-based neural network architecture. We present experimental results of the proposed model on a real-world patent dataset covering patent applications in diverse domains in which our organization is carrying out research and innovation activities, and also three publicly available sentence classification datasets. Through empirical analysis, we show that a) Our model performs significantly better than several strong baselines on the patent dataset; b) Performing dilation operation on LSTMs allows us to capture long term dependencies; c) our model is comparable to existing state-of-art approaches on the publicly available datasets; d) Error analysis through LIME shows that the proposed approach can help patent attorneys to interpret the decisions taken by the classifier.

[1]  Xiaoyan Zhu,et al.  Sentiment Analysis by Capsules , 2018, WWW.

[2]  Daniel P. W. Ellis,et al.  Feed-Forward Networks with Attention Can Solve Some Long-Term Memory Problems , 2015, ArXiv.

[3]  Min Yang,et al.  Investigating Capsule Networks with Dynamic Routing for Text Classification , 2018, EMNLP.

[4]  Gong Ling,et al.  An improved TF-IDF approach for text classification , 2005 .

[5]  Franck Dernoncourt,et al.  Sequential Short-Text Classification with Recurrent and Convolutional Neural Networks , 2016, NAACL.

[6]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[7]  Yoon Kim,et al.  Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.

[8]  Laurent Romary,et al.  Multiple Retrieval Models and Regression Models for Prior Art Search , 2009, CLEF.

[9]  Saurabh Srivastava,et al.  Identifying Aggression and Toxicity in Comments using Capsule Network , 2018, TRAC@COLING 2018.

[10]  Francesco Romani,et al.  Ranking a stream of news , 2005, WWW '05.

[11]  Nello Cristianini,et al.  Classification using String Kernels , 2000 .

[12]  Lovekesh Vig,et al.  Hybrid BiLSTM-Siamese network for FAQ Assistance , 2017, CIKM.

[13]  Geoffrey E. Hinton,et al.  Dynamic Routing Between Capsules , 2017, NIPS.

[14]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[15]  John MacLaren Walsh,et al.  Improving Automated Patent Claim Parsing: Dataset, System, and Experiments , 2016, ArXiv.

[16]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[17]  Dennis McLeod,et al.  A Comparative Study for Email Classification , 2007 .

[18]  Kuldip K. Paliwal,et al.  Bidirectional recurrent neural networks , 1997, IEEE Trans. Signal Process..

[19]  Diyi Yang,et al.  Hierarchical Attention Networks for Document Classification , 2016, NAACL.

[20]  Jun Zhao,et al.  Recurrent Convolutional Neural Networks for Text Classification , 2015, AAAI.

[21]  Shoko Suzuki,et al.  Extraction of Keywords of Novelties From Patent Claims , 2016, COLING.

[22]  Lovekesh Vig,et al.  Hierarchical Capsule Based Neural Network Architecture for Sequence Labeling , 2019, 2019 International Joint Conference on Neural Networks (IJCNN).

[23]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[24]  Zhiyuan Liu,et al.  A C-LSTM Neural Network for Text Classification , 2015, ArXiv.

[25]  Prerna Khurana,et al.  Detecting Aggression and Toxicity using a Multi Dimension Capsule Network , 2019, Proceedings of the Third Workshop on Abusive Language Online.

[26]  Dan Roth,et al.  Learning Question Classifiers , 2002, COLING.

[27]  Björn W. Schuller,et al.  Contextual Bidirectional Long Short-Term Memory Recurrent Neural Network Language Models: A Generative Approach to Sentiment Analysis , 2017, EACL.

[28]  Alex Graves,et al.  Neural Machine Translation in Linear Time , 2016, ArXiv.