Entity Alignment Through Joint Utilization of Multiple Pretrained Models for Attribution Relationship

Most existing models for entity alignment tasks are based on a single pre-training mode. Through multiple experiments, it has been verified that a single pre-training model is no longer sufficient to meet the needs of current entity alignment tasks. Based on considerations of multiple pre-training m...

Full description

Saved in:
Bibliographic Details
Published inIEEE International Conference on Cyber Security and Cloud Computing (CSCloud) (Online) pp. 1 - 6
Main Authors Yujia, Liu, Tao, Xu, Zhehan, Su, Xinsong, Li, Kaipeng, Xue, Yuxin, Liu
Format Conference Proceeding
LanguageEnglish
Published IEEE 28.06.2024
Subjects
Online AccessGet full text
ISSN2693-8928
DOI10.1109/CSCloud62866.2024.00008

Cover

Loading…
Abstract Most existing models for entity alignment tasks are based on a single pre-training mode. Through multiple experiments, it has been verified that a single pre-training model is no longer sufficient to meet the needs of current entity alignment tasks. Based on considerations of multiple pre-training models, this paper proposes an entity alignment method that utilizes multiple pre-training models. The purpose of this method is to address the limited accuracy in entity alignment tasks, with a particular focus on separately pre-training the embedding of entity attribute information and link relationship information for matching. To achieve the goal of improving entity alignment accuracy, this paper adopts joint alignment using pre-training with BERT and GloVe word embeddings. Experimental evaluations are conducted on the DBP15k dataset, and the results demonstrate that leveraging multiple pre-training models for attribute and relationship joint alignment can enhance model performance. The evaluation metrics of the model on three different language datasets, Hits@1 and MRR, are as follows: 0.899, 0.930, 0.964, and 0.932, 0.951, 0.975, respectively. These metrics indicate that the model performs well in aligning entities and outperforms current mainstream entity alignment methods.
AbstractList Most existing models for entity alignment tasks are based on a single pre-training mode. Through multiple experiments, it has been verified that a single pre-training model is no longer sufficient to meet the needs of current entity alignment tasks. Based on considerations of multiple pre-training models, this paper proposes an entity alignment method that utilizes multiple pre-training models. The purpose of this method is to address the limited accuracy in entity alignment tasks, with a particular focus on separately pre-training the embedding of entity attribute information and link relationship information for matching. To achieve the goal of improving entity alignment accuracy, this paper adopts joint alignment using pre-training with BERT and GloVe word embeddings. Experimental evaluations are conducted on the DBP15k dataset, and the results demonstrate that leveraging multiple pre-training models for attribute and relationship joint alignment can enhance model performance. The evaluation metrics of the model on three different language datasets, Hits@1 and MRR, are as follows: 0.899, 0.930, 0.964, and 0.932, 0.951, 0.975, respectively. These metrics indicate that the model performs well in aligning entities and outperforms current mainstream entity alignment methods.
Author Yuxin, Liu
Tao, Xu
Zhehan, Su
Yujia, Liu
Kaipeng, Xue
Xinsong, Li
Author_xml – sequence: 1
  givenname: Liu
  surname: Yujia
  fullname: Yujia, Liu
  email: liuyujia001@126.com
  organization: Northwest Minzu University,Key Laboratory of Minzu Languages and Cultures Intelligent Information Processing,Lanzhou,Gansu
– sequence: 2
  givenname: Xu
  surname: Tao
  fullname: Tao, Xu
  email: alfredxly@163.com
  organization: Northwest Minzu University,Key Laboratory of Linguistic and Cultural Computing Ministry of Education,Lanzhou,Gansu
– sequence: 3
  givenname: Su
  surname: Zhehan
  fullname: Zhehan, Su
  email: suzhehan@foxmail.com
  organization: Northwest Minzu University,Key Laboratory of Minzu Languages and Cultures Intelligent Information Processing,Lanzhou,Gansu
– sequence: 4
  givenname: Li
  surname: Xinsong
  fullname: Xinsong, Li
  email: 2506415434@qq.com
  organization: Northwest Minzu University,Key Laboratory of Linguistic and Cultural Computing Ministry of Education,Lanzhou,Gansu
– sequence: 5
  givenname: Xue
  surname: Kaipeng
  fullname: Kaipeng, Xue
  email: 1324519045@qq.com
  organization: Northwest Minzu University,Key Laboratory of Minzu Languages and Cultures Intelligent Information Processing,Lanzhou,Gansu
– sequence: 6
  givenname: Liu
  surname: Yuxin
  fullname: Yuxin, Liu
  email: liuyx@shmtu.edu.cn
  organization: College of Information Engineering Shanghai Maritime University,Shanghai
BookMark eNotkMlOwzAYhA0CCSh9AyT8AileEi_HKiqbWoGgnCsn-d0auXblOIfy9ESFuYxG-mYOc4MuQgyA0D0lM0qJfqg_ax-HTjAlxIwRVs7IKHWGplpqxSvCpdCKnaNrJjQvlGbqCk37_nukOCNcy_IawSJkl4947t027CFkvN6lOGx3-DW6MX1l592PyS4GHC1eDT67gwf8niAn4wJ0eBU78D22MeF5zsk1w4n-AH-q9Tt3uEWX1vgepv8-QevHxbp-LpZvTy_1fFk4TXOhqZCMEqoaa1WlBDdQARdEtFa2otKgoGmkbbnktGrLzihlSs2tYMxK2Qk-QXd_sw4ANofk9iYdN5QIUtHxjl9BY1we
CODEN IEEPAD
ContentType Conference Proceeding
DBID 6IE
6IL
CBEJK
RIE
RIL
DOI 10.1109/CSCloud62866.2024.00008
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Proceedings Order Plan All Online (POP All Online) 1998-present by volume
IEEE Xplore All Conference Proceedings
IEEE Electronic Library (IEL)
IEEE Proceedings Order Plans (POP All) 1998-Present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
EISBN 9798350376982
EISSN 2693-8928
EndPage 6
ExternalDocumentID 10605150
Genre orig-research
GroupedDBID 6IE
6IF
6IL
6IN
AAWTH
ABLEC
ADZIZ
ALMA_UNASSIGNED_HOLDINGS
BEFXN
BFFAM
BGNUA
BKEBE
BPEOZ
CBEJK
CHZPO
IEGSK
OCL
RIE
RIL
ID FETCH-LOGICAL-i91t-916721018bff85863ae5e3606cf7c659e8ebb7fc37315c4da88a493f622f77d63
IEDL.DBID RIE
IngestDate Wed Aug 27 02:35:15 EDT 2025
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-i91t-916721018bff85863ae5e3606cf7c659e8ebb7fc37315c4da88a493f622f77d63
PageCount 6
ParticipantIDs ieee_primary_10605150
PublicationCentury 2000
PublicationDate 2024-June-28
PublicationDateYYYYMMDD 2024-06-28
PublicationDate_xml – month: 06
  year: 2024
  text: 2024-June-28
  day: 28
PublicationDecade 2020
PublicationTitle IEEE International Conference on Cyber Security and Cloud Computing (CSCloud) (Online)
PublicationTitleAbbrev CSCLOUD
PublicationYear 2024
Publisher IEEE
Publisher_xml – name: IEEE
SSID ssj0003203974
Score 1.876912
Snippet Most existing models for entity alignment tasks are based on a single pre-training mode. Through multiple experiments, it has been verified that a single...
SourceID ieee
SourceType Publisher
StartPage 1
SubjectTerms Accuracy
BERT
Cloud computing
Computational modeling
Deep learning
entity alignment
GloVe
joint alignment
knowledge graph
Measurement
Semantics
Training
Title Entity Alignment Through Joint Utilization of Multiple Pretrained Models for Attribution Relationship
URI https://ieeexplore.ieee.org/document/10605150
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3PS8MwFA66kycVJ_4mB6-da5M06XGMjTHYENxgt9E0L1oc69D2oH-9L2lXURC8JTmEkPDy8vK-732E3EulpeYCAmvQ3LiBNNBMoF0lCq9GZXUILqM7m8eTJZ-uxKohq3suDAB48Bn0XNPn8k2RVe6rDC08dpIkGKEf4jw1Wav9UGFRH30rbzBcYT95GD4NN0VlHPnSgREi7osVqh86Kt6NjI_JfL-AGj3y2qtK3cs-f9Vm_PcKT0j3m7FHH1tfdEoOYHtGYOQ4uB90sMmffdKfLmpVHjotcuwty3zT0DBpYemswRbiROCVI8BQJ5W2eaf4sqWDshXHoi2E7iXfdcliPFoMJ0EjqxDkSVji7RZj1NcPlbZWCRWzFAQwjGMyK7NYJKBAa2kzJlkoMm5SpVKeMBtHkZXSxOycdLbFFi4ITSNrNdotuCJzJhI6ybh0JeGFtfiyspek67ZovasLZ6z3u3P1x_g1OXLH5JBYkbohnfKtglv0-aW-82f9BdiHrbE
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3PS8MwFA4yD3pSceJvc_DauTZJkx7H2JhzG4Id7Daa5kWLYx3aHvSvN0m7ioLgLc2hhISX917e970PoVsuJJeUgaeVMTeqIPEkYcauImGuRqGlD7aiO52FozkdL9iiJqs7LgwAOPAZdOzQ1fJVnpb2qcxYeGglSUyGvmscP40qulbzpEKCrvGutEZx-d3orv_UX-WlsvRLC0cIqGtXKH4oqThHMjxAs-0SKvzIa6csZCf9_NWd8d9rPETtb84efmy80RHagfUxgoFl4X7g3ip7dmV_HFe6PHicZ-ZrXmSrmoiJc42nNbrQ_AicdgQobMXSVu_YxLa4VzTyWLgB0b1kmzaKh4O4P_JqYQUvi_zC3G-hyfu6vpBaCyZCkgADYjKZVPM0ZBEIkJLrlHDis5SqRIiERkSHQaA5VyE5Qa11voZThJNAa2ksF2ybORUwGaWU26bwTGsTW-kz1LZbtNxUrTOW2905_2P-Bu2N4ulkObmfPVygfXtkFpcViEvUKt5KuDIRQCGv3bl_ASNdsQE
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=IEEE+International+Conference+on+Cyber+Security+and+Cloud+Computing+%28CSCloud%29+%28Online%29&rft.atitle=Entity+Alignment+Through+Joint+Utilization+of+Multiple+Pretrained+Models+for+Attribution+Relationship&rft.au=Yujia%2C+Liu&rft.au=Tao%2C+Xu&rft.au=Zhehan%2C+Su&rft.au=Xinsong%2C+Li&rft.date=2024-06-28&rft.pub=IEEE&rft.eissn=2693-8928&rft.spage=1&rft.epage=6&rft_id=info:doi/10.1109%2FCSCloud62866.2024.00008&rft.externalDocID=10605150