[1]刘宇昊,高 榕,严灵毓,等.基于图对比学习的长文本分类模型[J].湖北工业大学学报,2023,(5):67-74.
 LIU Yuhao,GAO Rong,YAN Lingyu,et al.A Long Text Classification Model Based on Graph Contrast Learning[J].,2023,(5):67-74.
点击复制

基于图对比学习的长文本分类模型()
分享到:

《湖北工业大学学报》[ISSN:1003-4684/CN:42-1752/Z]

卷:
期数:
2023年第5期
页码:
67-74
栏目:
出版日期:
2023-10-30

文章信息/Info

Title:
A Long Text Classification Model Based on Graph Contrast Learning
文章编号:
1003-4684(2023)05-0067-08
作者:
刘宇昊高 榕严灵毓叶志伟
湖北工业大学计算机学院,湖北 武汉 430068
Author(s):
LIU YuhaoGAO RongYAN LingyuYE Zhiwei
School of Computer Science,Hubei Univ. of Tech .,Wuhan 430068,China
关键词:
文本表示长文本分类图对比学习负采样
Keywords:
text representationlong text classificationgraph contrastive learningnegative sampling
分类号:
TP391.1
文献标志码:
A
摘要:
当前基于字符级考虑的文本分类方法在长文本分类上,存在输入维度过大致使计算困难以及内容过长难以捕捉长距离关系,从而导致准确度不足的问题.由此,提出基于自适应视图生成器和负采样优化的图对比学习长文本分类模型.首先将长文本分为若干段落,用 BERT衍生模型对段落进行嵌入表示,然后基于文本的高级结构将段落的嵌入表示视为节点构建图模型,接着使用自适应视图生成器对图进行增广,并通过图对比学习得到文本的嵌入表示,同时在图对比学习的负采样阶段,引入 PULearning知识修正负采样偏差的问题,最后将得到的文本嵌入表示使用两层线性层进行分类.通过在两个中文数据集上的实验显示,方法优于主流先进模型.
Abstract:
The current text classification methods based on characterlevel consideration have the problems of computational difficulty due to the large input dimension and the difficulty of capturing the longdistance relationship due to the long content, which leads to a lack of accuracy in long text classification. Thus, the proposed graph contrast learning long text classification model is based on an adaptive view generator and negative sampling optimization. Specifically, the long text is first divided into several paragraphs, and the paragraphs are embedded with the BERTderived model, then the graph model is constructed based on the highlevel structure of the text by considering the embedded representation of the paragraphs as nodes, then the graph is augmented using the adaptive view generator, and the embedded representation of the text is obtained by graph contrast learning, while PU learning knowledge is introduced to alleviate the problem of negative sampling bias in the negative sampling phase of graph contrast learning, and finally the obtained embedded representation of the text is classified using two linear layers. Experiments on two Chinese datasets show that the method outperforms mainstream advanced models.

参考文献/References:

[1]  KOWSARI K,JAFARI MEIMANDI K,HEIDARG YSAFA M,etal.Textclassificationalgorithms:AsurG vey[J].Information,2019,10(04):150. [2] MEDHAT W,HASSAN A,KORASHY H.SenGtiment analysisalgorithmsandappGlications:Asurvey[J].Ain Shamsengineeringjournal,2014,5(04):1093G1113. [3] YATES A,NOGUEIRA R,LIN J.PretrainedtransG formersfortextranking:BERTandbeyond[C]∥ProG ceedingsofthe14thACMInternationalConferenceon WebSearchandDataMining.2021:1154G1156. [4]  MA X,ZHU Q,ZHOU Y,etal.Improvingquestion generation withsentenceGlevelsemantic matchingand answerpositioninferring[C]∥ Proceedings ofthe AAAIConferenceon ArtificialIntelligence.2020,34 (05):8464G8471. [5]  WANG Z,LIU X,YANG P,etal.CrossGlingualtext classificationwithheterogeneousgraphneuralnetwork [C]∥ Proceedingsofthe59thAnnualMeetingofthe AssociationforComputationalLinguisticsandthe11th InternationalJoint Conference on Natural Language Processing,(ACL/IJCNLP)2021,(Volume2:Short Papers),2021:612G620 [6] CHAFFARS,INKPEND.UsingaheterogeneousdataG setforemotionanalysisintext[C]∥CanadianconferG enceonartificialintelligence.Springer,Berlin,HeidelG berg,2011:62G67. [7] MIKOLOVT,CHEN K,CORRADO G,etal.Efficient estimationofwordrepresentationsinvectorspace[C] ∥ 1stInternationalConferenceonLearningRepresenG tations,(ICLR2013),Scottsdale,Arizona,USA,2013. http:∥arxiv.org/abs/1301.3781. [8] LILLEBERGJ,ZHU Y,ZHANG Y.Supportvector machinesandword2vecfortextclassificationwithseG manticfeatures[C]∥2015IEEE 14thInternational ConferenceonCognitiveInformatics& CognitiveComG puting(ICCI? CC).IEEE,2015:136G140. [9] VASWANIA,SHAZEER N,PARMAR N,etal.AtG tentionisallyouneed[C]∥ AdvancesinNeuralInforG mationProcessingSystems30:AnnualConferenceon NeuralInformation Processing Systems 2017.2017: 5998G6008. [10]DEVLINJ,CHANG M W,LEEK,etal.Bert:PreGtrainG ingofdeepbidirectionaltransformersforlanguageunG derstanding[C]∥ProceedingsofNAACLGHLT.2019: 4171G4186. [11]VAN DEN OORD A,LIY,VinyalsO.Representation learningwithcontrastivepredictivecoding[C].CoRR, 2018,abs/1807.03748.http:∥ arxiv.org/abs/1807. 03748. [12]SUNC,QIU X,XU Y,etal.HowtofineGtunebertfor textclassification? [J]∥Chinanationalconferenceon Chinese computational linguistics.Springer,Cham, 2019:194G206. [13]MOHANTYI,GOYAL A,DOTTERWEICH A.EmoG tionsaresubtle:learningsentimentbasedtextrepreG sentationsusingcontrastivelearning[J/OL].[2021G04G  第38卷第5期            刘宇昊,等 基于图对比学习的长文本分类模型 73 15].CoRR,2021,abs/2112.01054.https:∥arxiv.org/ abs/2112.01054. [14]XU P,CHEN X,MA X ,etal.ContrastiveDocument Representation Learning with Graph Attention NetG works[C]∥ FindingsoftheAssociationforComputaG tionalLinguistics:EMNLP2021:3874–3884. [15]DUPLESSIS M,NIU G,SUGIYAMA M.ConvexforG mulationforlearningfrompositiveandunlabeleddata [C]∥Internationalconferenceon machinelearning. PMLR,2015:1386G1394. [16]VELIC ? KOVIC ' P,CUCURULLG,CASANOVA A,et al.Graphattentionnetworks[J].stat,2017,1050:20. [17]JANGE,GUS,POOLEB.CategoricalreparameterizaG tionwithgumbelGsoftmax[C]∥5THInternationalconG ferenceonlearningrepresentations,(ICLR2017),TouG lon,France,2018.https:∥openreview.net/forum? id= rke3y85ee. [18]CHU G,WANGX,SHIC,etal.CuCo:GraphrepresenG tationwithcurriculumcontrastivelearning[C]∥Proc. IJCAI.2021:2300G2306. [19]KIRYOR,NIUG,DUPLESSISMC,etal.PositiveGunG labeledlearningwithnonGnegativeriskestimator[C]∥ Advancesin NeuralInformation Processing Systems 30:AnnualConferenceonNeuralInformationProcessG ingSystems2017.2017:1675G1685. [20] HINTON G,VINYALS O,DEAN J.Distillingthe knowledgeinaneuralnetwork[J/OL].[2022G04G15]. CoRR,2015,abs/1503.02531.http:∥arxiv.org/abs/ 1503.02531. [21]WANGF,LIU H.UnderstandingthebehaviourofcontG rastiveloss[C]∥ProceedingsoftheIEEE/CVFconferG enceoncomputervisionandpatternrecognition.2021: 2495G2504. [22]李 景 阳,孙 茂 松.NonGindependentterm selectionfor Chinesetextcategorization[J].TsinghuaScienceand Technology,2009(01):115G122. [23]WANG C,ZHANG M,MA S,etal.Automaticonline newsissueconstructioninwebenvironment[C]∥ ProG ceedingsofthe17thInternationalConferenceonWorld WideWeb,WWW ’08,pages457G466,NewYork,NY, USA,2008.ACM. [24]LAIS,XUL,LIUK,etal.RecurrentconvolutionalneuG ralnetworksfortextclassification[C]∥TwentyGninth AAAIconferenceonartificialintelligence.2015. [25]ZHOU P,SHIW,TIANJ,etal.AttentionGbasedbidiG rectionallongshortGterm memorynetworksforrelation classification[C]∥ Proceedings ofthe54th annual meetingoftheassociationforcomputationallinguistics (volume2:Shortpapers).2016:207G212. [26]KIMJ,JANGS,PARKE,etal.Textclassificationusing capsules[J].Neurocomputing,2020,376:214G221. [27]BELTAGYI,PETERS M E,COHAN A.Longformer: ThelongGdocumenttransformer[J/OL].[2022G04G15]. CoRR,2020,abs/2004.05150.https:∥arxiv.org/abs/ 2004.05150. [28]HUAN H,YANJ,XIEY,etal.FeatureGenhancednonG equilibriumbidirectionallongshortGterm memorymodG elfor Chinesetextclassification[J].IEEE Access, 2020,8:199629G199637. [29]YANGZ,YANG D,DYERC,etal.HierarchicalattenG tionnetworksfordocumentclassification[C]∥ProG ceedingsofthe2016conferenceoftheNorthAmerican chapteroftheassociationforcomputationallinguistics: humanlanguagetechnologies.2016:1480G1489. [30]YOU Y,CHEN T,SUI Y,etal.Graphcontrastive learningwithaugmentations[J].AdvancesinNeuralInG formationProcessingSystems,2020,33:5812G5823.

备注/Memo

备注/Memo:
[收稿日期]2022 06 12 [第一作者]刘宇昊(1996-),男,湖北咸宁人,湖北工业大学硕士研究生,研究方向为电子信息.
更新日期/Last Update: 2023-10-26