TRIZ-RAGNER: A Retrieval-Augmented Large Language Model for TRIZ-Aware Named Entity Recognition in Patent-Based Contradiction Mining

Authors

  • Zitong Xu Sun Yat sen University, Guangzhou, Guangdong, China Author
  • Yuqing Wu Uber Technologies, Inc., Seattle, WA, USA Author
  • Yue Zhao Monroe University, New Rochelle, NY, USA Author

DOI:

https://doi.org/10.71222/8x5p1n51

Keywords:

TRIZ contradiction mining, named entity recognition, large language models, retrieval-augmented generation, patent analysis

Abstract

TRIZ-based contradiction mining is a fundamental task in patent analysis and systematic innovation, as it enables the identification of improving and worsening technical parameters that drive inventive problem solving. However, existing approaches largely rely on rule-based systems or traditional machine learning models, which struggle with semantic ambiguity, domain dependency, and limited generalization when processing complex patent language. Recently, large language models (LLMs) have shown strong semantic understanding capabilities, yet their direct application to TRIZ parameter extraction remains challenging due to hallucination and insufficient grounding in structured TRIZ knowledge. To address these limitations, this paper proposes TRIZ-RAGNER, a retrieval-augmented large language model framework for TRIZ-aware named entity recognition in patent-based contradiction mining. TRIZ-RAGNER reformulates contradiction mining as a semantic-level NER task and integrates dense retrieval over a TRIZ knowledge base, cross-encoder reranking for context refinement, and structured LLM prompting to extract improving and worsening parameters from patent sentences. By injecting domain-specific TRIZ knowledge into the LLM reasoning process, the proposed framework effectively reduces semantic noise and improves extraction consistency. Experiments on the PaTRIZ dataset demonstrate that TRIZ-RAGNER consistently outperforms traditional sequence labeling models and LLM-based baselines. The proposed framework achieves a precision of 85.6%, a recall of 82.9%, and an F1-score of 84.2% in TRIZ contradiction pair identification. Compared with the strongest baseline using prompt-enhanced GPT, TRIZ-RAGNER yields an absolute F1-score improvement of 7.3 percentage points, confirming the effectiveness of retrieval-augmented TRIZ knowledge grounding for robust and accurate patent-based contradiction mining.

References

1. G. Guarino, "Text mining for automating TRIZ-based inventive design process using patent documents (Doctoral dissertation, Universite de Strasbourg)," 2022.

2. D. Cavallucci, and N. Khomenko, "From TRIZ to OTSM-TRIZ: addressing complexity challenges in inventive design," International Journal of Product Development, vol. 4, no. 1-2, pp. 4-21, 2007.

3. G. Guarino, A. Samet, and D. Cavallucci, "PaTRIZ: A framework for mining TRIZ contradictions in patents," Expert Systems with Applications, vol. 207, p. 117942, 2022. doi: 10.1016/j.eswa.2022.117942

4. P. R. Rao, S. L. Devi, and P. Rosso, "Automatic identification of concepts and conceptual relations from patents using machine learning methods," Icon, vol. 2013, pp. 18-20, 2013.

5. A. Akhundov, D. Trautmann, and G. Groh, "Sequence labeling: A practical approach," arXiv preprint arXiv:1808.03926, 2018.

6. J. Devlin, M. W. Chang, K. Lee, and K. Toutanova, "Bert: Pre-training of deep bidirectional transformers for language understanding," In Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, volume 1 (long and short papers), June, 2019, pp. 4171-4186.

7. J. Lee, W. Yoon, S. Kim, D. Kim, S. Kim, C. H. So, and J. Kang, "BioBERT: a pre-trained biomedical language representation model for biomedical text mining," Bioinformatics, vol. 36, no. 4, pp. 1234-1240, 2020. doi: 10.1093/bioinformatics/btz682

8. I. Beltagy, K. Lo, and A. Cohan, "SciBERT: A pretrained language model for scientific text," In Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP), November, 2019, pp. 3615-3620.

9. T. Brown, B. Mann, N. Ryder, M. Subbiah, J. D. Kaplan, P. Dhariwal, and D. Amodei, "Language models are few-shot learners," Advances in neural information processing systems, vol. 33, pp. 1877-1901, 2020.

10. J. Wei, X. Wang, D. Schuurmans, M. Bosma, F. Xia, E. Chi, and D. Zhou, "Chain-of-thought prompting elicits reasoning in large language models," Advances in neural information processing systems, vol. 35, pp. 24824-24837, 2022. doi: 10.52202/068431-1800

11. Z. Liang, W. Wei, K. Zhang, and H. Chen, "Research on multi-hop inference optimization of llm based on mquake framework," arXiv preprint arXiv:2509.04770, 2025.

12. L. Jiang, and S. M. Goetz, "Natural language processing in the patent domain: a survey," Artificial Intelligence Review, vol. 58, no. 7, p. 214, 2025. doi: 10.1007/s10462-025-11168-z

13. P. Lewis, E. Perez, A. Piktus, F. Petroni, V. Karpukhin, N. Goyal, and D. Kiela, "Retrieval-augmented generation for knowledge-intensive nlp tasks," Advances in neural information processing systems, vol. 33, pp. 9459-9474, 2020.

Downloads

Published

04 March 2026

Issue

Section

Article

How to Cite

Xu, Z., Wu, Y., & Zhao, Y. (2026). TRIZ-RAGNER: A Retrieval-Augmented Large Language Model for TRIZ-Aware Named Entity Recognition in Patent-Based Contradiction Mining. Journal of Computer, Signal, and System Research, 3(2), 56-65. https://doi.org/10.71222/8x5p1n51