快速检索:      
引用本文:
【打印本页】   【下载PDF全文】   查看/发表评论  【EndNote】   【RefMan】   【BibTex】
←前一篇|后一篇→ 过刊浏览    高级检索
本文已被:浏览 138次   下载 231 本文二维码信息
码上扫一扫!
分享到: 微信 更多
基于BERT与细粒度特征提取的数据法学问答系统
宋文豪1, 汪洋1, 朱苏磊1, 张倩1, 吴晓燕2
1.上海师范大学 信息与机电工程学院, 上海 201418;2.上海交通大学 电子信息与电气工程学院, 上海 200240
摘要:
首先利用bidirectional encoder representations from transformers(BERT)模型的强大的语境理解能力来提取数据法律文本的深层语义特征,然后引入细粒度特征提取层,依照注意力机制,重点关注文本中与数据法律问答相关的关键部分,最后对所采集的法律问答数据集进行训练和评估. 结果显示:与传统的多个单一模型相比,所提出的模型在准确度、精确度、召回率、F1分数等关键性能指标上均有提升,表明该系统能够更有效地理解和回应复杂的数据法学问题,为研究数据法学的专业人士和公众用户提供更高质量的问答服务.
关键词:  bidirectional encoder representations from transformers(BERT)模型  细粒度特征提取  注意力机制  自然语言处理(NLP)
DOI:10.3969/J.ISSN.1000-5137.2024.02.010
分类号:TP3911
基金项目:上海市科学仪器领域项目(22142201900);教育部重大项目(20JZD020);国家自然科学基金(62301320)
Data law Q&A system based on BERT and fine-grained feature extraction
SONG Wenhao1, WANG Yang1, ZHU Sulei1, ZHANG Qian1, WU Xiaoyan2
1.College of Information, Mechanical and Electrical Engineering, Shanghai Normal University, Shanghai 201418, China;2.School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
Abstract:
A data legal question and answer system was proposed based on bidirectional encoder representations from transformers(BERT) model and fine-grained feature extraction to provide accurate professional legal consulting services. Firstly, the powerful contextual understanding ability of the BERT model was leveraged to extract deep semantic features from data legal texts. Subsequently, a fine-grained feature extraction layer was introduced which mainly focused on key components related to data legal Q&A within the text using an attention mechanism. Finally, the collected legal Q&A dataset was trained and evaluated. The results indicated that compared to traditional multiple single models, the proposed model showed improvements in key performance indicators such as accuracy, precision, recall, and F1 score, which suggested that the system could more effectively comprehend and address complex issues in data law, providing higher quality Q&A services for both research data law professionals and the general public.
Key words:  bidirectional encoder representations from transformers(BERT) model  fine-grained feature extraction  attention mechanism  natural language processing (NLP)