<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">IASC</journal-id>
<journal-id journal-id-type="nlm-ta">IASC</journal-id>
<journal-id journal-id-type="publisher-id">IASC</journal-id>
<journal-title-group>
<journal-title>Intelligent Automation &#x0026; Soft Computing</journal-title>
</journal-title-group>
<issn pub-type="epub">2326-005X</issn>
<issn pub-type="ppub">1079-8587</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">20218</article-id>
<article-id pub-id-type="doi">10.32604/iasc.2022.020218</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>MSM: A Method of Multi-Neighborhood Sampling Matching for Entity Alignment</article-title><alt-title alt-title-type="left-running-head">MSM: A Method of Multi-neighborhood Sampling Matching for Entity Alignment</alt-title><alt-title alt-title-type="right-running-head">MSM: A Method of Multi-neighborhood Sampling Matching for Entity Alignment</alt-title>
</title-group>
<contrib-group content-type="authors">
<contrib id="author-1" contrib-type="author">
<name name-style="western"><surname>Lu</surname><given-names>Donglei</given-names></name>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
<contrib id="author-2" contrib-type="author">
<name name-style="western"><surname>Sun</surname><given-names>Yundong</given-names></name>
<xref ref-type="aff" rid="aff-2">2</xref>
</contrib>
<contrib id="author-3" contrib-type="author">
<name name-style="western"><surname>Dai</surname><given-names>Qinrui</given-names></name>
<xref ref-type="aff" rid="aff-2">2</xref>
</contrib>
<contrib id="author-4" contrib-type="author" corresp="yes">
<name name-style="western"><surname>Li</surname><given-names>Xiaofang</given-names></name>
<xref ref-type="aff" rid="aff-3">3</xref><email>lixiaofang@hit.edu.cn</email>
</contrib>
<contrib id="author-5" contrib-type="author">
<name name-style="western"><surname>Zhu</surname><given-names>Dongjie</given-names></name>
<xref ref-type="aff" rid="aff-4">4</xref>
</contrib>
<contrib id="author-6" contrib-type="author">
<name name-style="western"><surname>Du</surname><given-names>Haiwen</given-names></name>
<xref ref-type="aff" rid="aff-2">2</xref>
</contrib>
<contrib id="author-7" contrib-type="author">
<name name-style="western"><surname>Wang</surname><given-names>Yansong</given-names></name>
<xref ref-type="aff" rid="aff-4">4</xref>
</contrib>
<contrib id="author-8" contrib-type="author">
<name name-style="western"><surname>Qu</surname><given-names>Rongning</given-names></name>
<xref ref-type="aff" rid="aff-3">3</xref>
</contrib>
<contrib id="author-9" contrib-type="author">
<name name-style="western"><surname>Cao</surname><given-names>Ning</given-names></name>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
<contrib id="author-10" contrib-type="author">
<name name-style="western"><surname>O&#x2019;Hare</surname><given-names>Gregory M. P.</given-names></name>
<xref ref-type="aff" rid="aff-5">5</xref>
</contrib>
<aff id="aff-1"><label>1</label><institution>School of Artificial Intelligence, Wuxi Vocational College of Science and Technology</institution>, <addr-line>Wuxi, 214028</addr-line>, <country>China</country></aff>
<aff id="aff-2"><label>2</label><institution>School of Astronautics, Harbin Institute of Technology</institution>, <addr-line>Harbin, 150001</addr-line>, <country>China</country></aff>
<aff id="aff-3"><label>3</label><institution>Department of Mathematics, Harbin Institute of Technology</institution>, <addr-line>Weihai, 264209</addr-line>, <country>China</country></aff>
<aff id="aff-4"><label>4</label><institution>School of Computer Science and Technology, Harbin Institute of Technology</institution>, <addr-line>Weihai, 264209</addr-line>, <country>China</country></aff>
<aff id="aff-5"><label>5</label><institution>School of Computer Science, University College Dublin</institution>, <addr-line>Dublin, Dublin4</addr-line>, <country>Ireland</country></aff>
</contrib-group><author-notes><corresp id="cor1">&#x002A;Corresponding Author: Xiaofang Li. Email: <email>lixiaofang@hit.edu.cn</email></corresp></author-notes>
<pub-date pub-type="epub" date-type="pub" iso-8601-date="2021-10-19"><day>19</day>
<month>10</month>
<year>2021</year></pub-date>
<volume>32</volume>
<issue>2</issue>
<fpage>1141</fpage>
<lpage>1151</lpage>
<history>
<date date-type="received"><day>14</day><month>5</month><year>2021</year></date>
<date date-type="accepted"><day>15</day><month>6</month><year>2021</year></date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2021 Lu et al.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Lu et al.</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_IASC_20218.pdf"></self-uri>
<abstract>
<p>The heterogeneity of knowledge graphs brings great challenges to entity alignment. In particular, the attributes of network entities in the real world are complex and changeable. The key to solving this problem is to expand the neighborhoods in different ranges and extract the neighborhood information efficiently. Based on this idea, we propose Multi-neighborhood Sampling Matching Network (MSM), a new KG alignment network, aiming at the structural heterogeneity challenge. MSM constructs a multi-neighborhood network representation learning method to learn the KG structure embedding. It then adopts a unique sampling and cosine cross-matching method to solve different sizes of neighborhoods and distinct topological structures in two entities. To choose the right neighbors, we apply a down-sampling process to select the most informative entities towards the central target entity from its one-hop and two-hop neighbors. To verify the effectiveness of matching this neighborhood with any neighborhood in the corresponding node, we give a cosine cross-graph neighborhood matching method and conduct detailed research and analysis on three entity matching datasets, which proves the effectiveness of MSM.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Entity alignment</kwd>
<kwd>representation learning</kwd>
<kwd>heterogeneous network</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Introduction</title>
<p>Entity alignment is designed to determine whether two or more entities with different knowledge graphs point to the same object in the real world. It not only contributes to the construction and expansion of knowledge base, but also plays an important role in solving cross-network crimes. For example, entity alignment is now widely used in graph networks and social networks [<xref ref-type="bibr" rid="ref-1">1</xref>,<xref ref-type="bibr" rid="ref-2">2</xref>]. The most advanced entity alignment solutions mainly rely on the structure information of knowledge map to judge the equivalence of entities, but in the real-world knowledge map, most entities only have low node degree and little structure information. In addition, the lack of annotated data greatly limits the effectiveness of the entity alignment model. Unfortunately, entity alignment is not trivial, because real-life knowledge graphs are often incomplete and different knowledge graphs typically have heterogeneous schemas and the equivalent entities in different graphs often have different neighborhood structures.</p>
<p>The aim of network entity alignment technique is to find the same entity in different networks. At present, increasing attention has been paid to the embedding-based methods [<xref ref-type="bibr" rid="ref-3">3</xref>&#x2013;<xref ref-type="bibr" rid="ref-7">7</xref>]. Its core idea is to transform the nodes in the network into low-dimensional vectors by certain methods, whose vector dimensions are much lower than the number of nodes and contain the attributes of the nodes themselves and the semantic relationship information between the nodes. Unfortunately, these methods can only deal with homogeneous networks or multilingual knowledge graphs. The types of nodes and the relationships between nodes are same [<xref ref-type="bibr" rid="ref-8">8</xref>,<xref ref-type="bibr" rid="ref-9">9</xref>]. However, most of the network data in reality are heterogeneous networks. There are many kinds of node types or multiple node relationships. For example, in <xref ref-type="fig" rid="fig-1">Fig. 1</xref>, <italic>United States of America</italic> is among the one-hop (direct) neighbors of <italic>Barack Obama</italic> in Wikidata. However, in Dbpedia, it is a two-hop neighbors.</p>
<fig id="fig-1">
<label>Figure 1</label>
<caption>
<title>Non-isomorphic relational neighborhood of <italic>Barack Obama in</italic> DBpedia (left) and Wikidata (right), respectively</title></caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_20218-fig-1.png"/>
</fig>
<p>At the same time, the information contributed by the same order neighbor of the central entity is nonequivalent [<xref ref-type="bibr" rid="ref-8">8</xref>,<xref ref-type="bibr" rid="ref-9">9</xref>]. To elaborate on this point, let&#x2019;s move on to <xref ref-type="fig" rid="fig-1">Fig. 1</xref>. There are many city entities for USA which also have the entity <italic>United States of America</italic>. Consequently, the contribution of <italic>United States of America</italic> to the central entity is significantly less than that of <italic>Christian</italic>. Because existing embedding-based method are unable to choose the right neighbors, we need a better approach.</p>
<p>The challenge of resolving this issue lies in the difficulty of fully mitigating the non-isomorphism in the neighborhood structures of counterpart entities from different KGs. We present the Multi-neighborhood Sampling Matching Network (MSM), a novel Network embedded and Sampling-based framework. The goal of MSM is to obtain the neighborhood with the most valuable information and accurately estimate the similarity of neighborhood among entities in different knowledge graphs. To learn the embedding of knowledge graph structure, MSM utilizes multi-neighborhood network representation learning method to aggregate higher degree neighboring structural information for entities. We evaluate MSM by applying it to benchmark datasets DBP15K [<xref ref-type="bibr" rid="ref-10">10</xref>&#x2013;<xref ref-type="bibr" rid="ref-16">16</xref>].</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Multi-Neighborhood Sampling Matching</title>
<sec id="s2_1">
<label>2.1</label>
<title>KG Structure Embedding</title>
<p>In addition to the one-hop neighborhood information, the non-directly related high-order neighborhood information is also very important for the representation of the central entity [<xref ref-type="bibr" rid="ref-16">16</xref>]. Therefore, MSM first constructed a multi-neighborhood network representation learning method (MNE), using this representation method learns the structure embedding of KG to gather the high-order neighborhood structure information of the entity. The overall architecture and processing pipeline of MNE is shown in <xref ref-type="fig" rid="fig-2">Fig. 2</xref>. MSM uses pre-trained word vectors to initialize MNE.</p>
<p>We use <inline-formula id="ieqn-1">
<mml:math id="mml-ieqn-1"><mml:mi>G</mml:mi><mml:mo>=</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mi>E</mml:mi><mml:mo>,</mml:mo><mml:mrow></mml:mrow><mml:mi>R</mml:mi><mml:mo>,</mml:mo><mml:mrow></mml:mrow><mml:mi>T</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math>
</inline-formula> to represent a knowledge graph, in which <inline-formula id="ieqn-2">
<mml:math id="mml-ieqn-2"><mml:mi>E</mml:mi><mml:mo>,</mml:mo><mml:mrow></mml:mrow><mml:mi>R</mml:mi><mml:mo>,</mml:mo><mml:mrow></mml:mrow><mml:mi>T</mml:mi></mml:math>
</inline-formula> represents entities, relationships, and sets of triples. Without loss of generality, we consider the entity alignment task between two knowledge graphs <inline-formula id="ieqn-3">
<mml:math id="mml-ieqn-3"><mml:mrow><mml:msub><mml:mi>G</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula> and <inline-formula id="ieqn-4">
<mml:math id="mml-ieqn-4"><mml:mrow><mml:msub><mml:mi>G</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula> based on a set of pre-aligned equivalent entities. Our goal is to find an equivalent entity pair between <inline-formula id="ieqn-5">
<mml:math id="mml-ieqn-5"><mml:mrow><mml:msub><mml:mi>G</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula> and <inline-formula id="ieqn-6">
<mml:math id="mml-ieqn-6"><mml:mrow><mml:msub><mml:mi>G</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula>. We put <inline-formula id="ieqn-7">
<mml:math id="mml-ieqn-7"><mml:mrow><mml:msub><mml:mi>G</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula> and <inline-formula id="ieqn-8">
<mml:math id="mml-ieqn-8"><mml:mrow><mml:msub><mml:mi>G</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula> as a large input graph into MSM, each MNE layer consists of two layers of GCN aggregated entities of one-hop neighborhood information, a layer of attention-network (ATT-Net) aggregated entities two-hop neighborhood information and each GNN updates the node representation as:</p>
<p><disp-formula id="eqn-1"><label>(1)</label>
<mml:math id="mml-eqn-1" display="block"><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mrow><mml:mi>Re</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">L</mml:mi><mml:mi mathvariant="normal">U</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:munder><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x222A;</mml:mo><mml:mo fence="false" stretchy="false">{</mml:mo><mml:mi>i</mml:mi><mml:mo fence="false" stretchy="false">}</mml:mo></mml:mrow></mml:munder><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:mfrac></mml:mrow><mml:msubsup><mml:mi>W</mml:mi><mml:mn>1</mml:mn><mml:mi>l</mml:mi></mml:msubsup></mml:mstyle></mml:mrow><mml:msubsup><mml:mi>h</mml:mi><mml:mi>j</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math>
</disp-formula>where <inline-formula id="ieqn-9">
<mml:math id="mml-ieqn-9"><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup></mml:math>
</inline-formula> is the output node features of <inline-formula id="ieqn-10">
<mml:math id="mml-ieqn-10"><mml:mi>l</mml:mi></mml:math>
</inline-formula>-th GCN layer. <inline-formula id="ieqn-11">
<mml:math id="mml-ieqn-11"><mml:mrow><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> is the normalization constant. <inline-formula id="ieqn-12">
<mml:math id="mml-ieqn-12"><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> is the set of neighbor indices of entity <inline-formula id="ieqn-13">
<mml:math id="mml-ieqn-13"><mml:mi>i</mml:mi></mml:math>
</inline-formula>. <inline-formula id="ieqn-14">
<mml:math id="mml-ieqn-14"><mml:msubsup><mml:mi>W</mml:mi><mml:mn>1</mml:mn><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mrow><mml:mrow><mml:msup><mml:mi>d</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msup></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:msup><mml:mi>d</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msup></mml:mrow></mml:mrow></mml:msup></mml:mrow></mml:math>
</inline-formula> is the attention weight of <inline-formula id="ieqn-15">
<mml:math id="mml-ieqn-15"><mml:mi>l</mml:mi></mml:math>
</inline-formula> layer neighborhood of node <inline-formula id="ieqn-16">
<mml:math id="mml-ieqn-16"><mml:mi>i</mml:mi></mml:math>
</inline-formula> obtained by GNN training. To aggregate the two-hop neighborhood information of entity <inline-formula id="ieqn-17">
<mml:math id="mml-ieqn-17"><mml:mi>i</mml:mi></mml:math>
</inline-formula>, we have adopted ATT-Net. The hidden representation of entity <inline-formula id="ieqn-18">
<mml:math id="mml-ieqn-18"><mml:mi>i</mml:mi></mml:math>
</inline-formula>, denoted as <inline-formula id="ieqn-19">
<mml:math id="mml-ieqn-19"><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mi>l</mml:mi></mml:msubsup></mml:math>
</inline-formula>, is computed as follows:</p>
<p><disp-formula id="eqn-2"><label>(2)</label>
<mml:math id="mml-eqn-2" display="block"><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mrow><mml:mi>Re</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">L</mml:mi><mml:mi mathvariant="normal">U</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:munder><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x222A;</mml:mo><mml:mo fence="false" stretchy="false">{</mml:mo><mml:mi>i</mml:mi><mml:mo fence="false" stretchy="false">}</mml:mo></mml:mrow></mml:munder><mml:mrow><mml:msubsup><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>W</mml:mi><mml:mn>2</mml:mn><mml:mi>l</mml:mi></mml:msubsup></mml:mrow><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>j</mml:mi><mml:mo>,</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math>
</disp-formula>where <inline-formula id="ieqn-20">
<mml:math id="mml-ieqn-20"><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup></mml:math>
</inline-formula> is the output node features of <inline-formula id="ieqn-21">
<mml:math id="mml-ieqn-21"><mml:mi>l</mml:mi></mml:math>
</inline-formula>-th ATT-Net layer. <inline-formula id="ieqn-22">
<mml:math id="mml-ieqn-22"><mml:msubsup><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup></mml:math>
</inline-formula> is a learnable attention weight for entities <inline-formula id="ieqn-23">
<mml:math id="mml-ieqn-23"><mml:mi>i</mml:mi></mml:math>
</inline-formula> and its neighbor <inline-formula id="ieqn-24">
<mml:math id="mml-ieqn-24"><mml:mi>j</mml:mi></mml:math>
</inline-formula>. <inline-formula id="ieqn-25">
<mml:math id="mml-ieqn-25"><mml:msubsup><mml:mi>W</mml:mi><mml:mn>2</mml:mn><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mrow><mml:mrow><mml:msup><mml:mi>d</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msup></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:msup><mml:mi>d</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msup></mml:mrow></mml:mrow></mml:msup></mml:mrow></mml:math>
</inline-formula> is the wight matrix. Finally, for entity <inline-formula id="ieqn-26">
<mml:math id="mml-ieqn-26"><mml:mi>i</mml:mi></mml:math>
</inline-formula>, we use the gating mechanism to combine its one-hop and two-hop neighborhood information. The node representations are given as below:</p>
<p><disp-formula id="eqn-3"><label>(3)</label>
<mml:math id="mml-eqn-3" display="block"><mml:mrow><mml:msubsup><mml:mi>h</mml:mi><mml:mi>i</mml:mi><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>g</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msubsup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>g</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msubsup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x22C5;</mml:mo><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msubsup><mml:mo>+</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mi>g</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msubsup></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x22C5;</mml:mo><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msubsup><mml:mo>+</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mi>g</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msubsup></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x22C5;</mml:mo><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msubsup></mml:mrow></mml:math>
</disp-formula>where <inline-formula id="ieqn-27">
<mml:math id="mml-ieqn-27"><mml:mi>g</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mi>&#x03C3;</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi>M</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup><mml:mo>+</mml:mo><mml:mrow><mml:msub><mml:mi>b</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math>
</inline-formula>, <inline-formula id="ieqn-28">
<mml:math id="mml-ieqn-28"><mml:mi>g</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mi>&#x03C3;</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi>M</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup><mml:mo>+</mml:mo><mml:mrow><mml:msub><mml:mi>b</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math>
</inline-formula> and <inline-formula id="ieqn-29">
<mml:math id="mml-ieqn-29"><mml:mrow><mml:msub><mml:mi>M</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula>, <inline-formula id="ieqn-30">
<mml:math id="mml-ieqn-30"><mml:mrow><mml:msub><mml:mi>M</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula> and <inline-formula id="ieqn-31">
<mml:math id="mml-ieqn-31"><mml:mrow><mml:msub><mml:mi>b</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula>, <inline-formula id="ieqn-32">
<mml:math id="mml-ieqn-32"><mml:mrow><mml:msub><mml:mi>b</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula> are the weight matrix and bias vector, respectively. In order to control the accumulated noise, we also introduce the highway network [<xref ref-type="bibr" rid="ref-15">15</xref>] into the MNE layer, which can effectively control the propagation of noise in the MNE layer.</p>
<fig id="fig-2">
<label>Figure 2</label>
<caption>
<title>Overall architecture and processing pipeline of MNE</title></caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_20218-fig-2.png"/>
</fig>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>Neighborhood Sampling</title>
<p>The first-order and second-order neighborhoods of an entity are the key to determining whether the entity should be aligned with other entities. However, as we discussed before, not all first-order and second-order neighborhoods make a positive contribution to entity alignment. In order to select the correct neighbor, we use a down-sampling process to select the entity that provides the most information to the central target entity from its first-order and second-order neighborhoods. Previously, many random sampling methods were used, but the method have some randomness. The neighbor nodes sampled may not necessarily contribute the most. Therefore, we give a sampling method based on probability. Formally, given an entity <inline-formula id="ieqn-33">
<mml:math id="mml-ieqn-33"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula>, the calculation formula for sampling its first-hop domain <inline-formula id="ieqn-34">
<mml:math id="mml-ieqn-34"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:math>
</inline-formula> and second-hop domain <inline-formula id="ieqn-35">
<mml:math id="mml-ieqn-35"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msup><mml:mi>j</mml:mi><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mrow></mml:msub></mml:mrow></mml:math>
</inline-formula> are [<xref ref-type="bibr" rid="ref-16">16</xref>]:</p>
<p><disp-formula id="eqn-4"><label>(4)</label>
<mml:math id="mml-eqn-4" display="block"><mml:mtable columnalign="right left" rowspacing=".5em" columnspacing="thickmathspace" displaystyle="true"><mml:mtr><mml:mtd><mml:mi>p</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mi>s</mml:mi><mml:mi>o</mml:mi><mml:mi>f</mml:mi><mml:mi>t</mml:mi><mml:mo movablelimits="true" form="prefix">max</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>W</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>j</mml:mi></mml:mrow><mml:mi>T</mml:mi></mml:msubsup><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mrow><mml:mi>exp</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>W</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>j</mml:mi></mml:mrow><mml:mi>T</mml:mi></mml:msubsup><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>&#x2208;</mml:mo><mml:msubsup><mml:mi>N</mml:mi><mml:mi>i</mml:mi><mml:mn>1</mml:mn></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mi>exp</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>W</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mi>T</mml:mi></mml:msubsup><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mi>p</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>j</mml:mi><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mi>s</mml:mi><mml:mi>o</mml:mi><mml:mi>f</mml:mi><mml:mi>t</mml:mi><mml:mo movablelimits="true" form="prefix">max</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>W</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>j</mml:mi><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow><mml:mi>T</mml:mi></mml:msubsup><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mrow><mml:mi>exp</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>W</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>j</mml:mi></mml:mrow><mml:mi>T</mml:mi></mml:msubsup><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>&#x2208;</mml:mo><mml:msubsup><mml:mi>N</mml:mi><mml:mi>i</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mi>exp</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>W</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:msubsup><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mi>T</mml:mi></mml:msubsup><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:mtd></mml:mtr></mml:mtable></mml:math>
</disp-formula></p>
<p>where <inline-formula id="ieqn-36">
<mml:math id="mml-ieqn-36"><mml:msubsup><mml:mi>N</mml:mi><mml:mi>i</mml:mi><mml:mn>1</mml:mn></mml:msubsup></mml:math>
</inline-formula> and <inline-formula id="ieqn-37">
<mml:math id="mml-ieqn-37"><mml:msubsup><mml:mi>N</mml:mi><mml:mi>i</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:math>
</inline-formula> are the first-hop and second-hop neighborhood index set of the central entity <inline-formula id="ieqn-38">
<mml:math id="mml-ieqn-38"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula>, respectively. <inline-formula id="ieqn-39">
<mml:math id="mml-ieqn-39"><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula>, <inline-formula id="ieqn-40">
<mml:math id="mml-ieqn-40"><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:math>
</inline-formula> and <inline-formula id="ieqn-41">
<mml:math id="mml-ieqn-41"><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msup><mml:mi>j</mml:mi><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mrow></mml:msub></mml:mrow></mml:math>
</inline-formula> are the entity representation vectors of the entities <inline-formula id="ieqn-42">
<mml:math id="mml-ieqn-42"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula>, <inline-formula id="ieqn-43">
<mml:math id="mml-ieqn-43"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:math>
</inline-formula> and <inline-formula id="ieqn-44">
<mml:math id="mml-ieqn-44"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msup><mml:mi>j</mml:mi><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mrow></mml:msub></mml:mrow></mml:math>
</inline-formula> respectively, and <inline-formula id="ieqn-45">
<mml:math id="mml-ieqn-45"><mml:mrow><mml:msub><mml:mi>W</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula>, <inline-formula id="ieqn-46">
<mml:math id="mml-ieqn-46"><mml:mrow><mml:msub><mml:mi>W</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula> are shared weight matrix.</p>
<p>By selectively sampling first-hop and second-hop neighborhoods, MSM gets the most valuable neighborhood information of each entity. In the end, MSM achieves the alignment of G1 and G2 through neighborhood matching and neighborhood aggregation.</p>
</sec>
<sec id="s2_3">
<label>2.3</label>
<title>Neighborhood Matching</title>
<p>In order to judge whether the two entities should be aligned, we need a similarity calculation to its neighbor nodes. MSM builds hence a neighborhood subgraph generated by the sampling process for each entity. Then MSM will only operate neighbors within the subgraph, achieving the neighborhood matching.</p>
<p>Inspired by the graph matching method [<xref ref-type="bibr" rid="ref-5">5</xref>], our specific matching method is as follows.</p>
<p>We need to compare the subgraphs of each candidate entity in its sampling neighborhood subgraph <inline-formula id="ieqn-47">
<mml:math id="mml-ieqn-47"><mml:mrow><mml:msub><mml:mi>E</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula> to select the best aligned entity of the entity <inline-formula id="ieqn-48">
<mml:math id="mml-ieqn-48"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> in <inline-formula id="ieqn-49">
<mml:math id="mml-ieqn-49"><mml:mrow><mml:msub><mml:mi>E</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula>. But if each entity neighborhood in <inline-formula id="ieqn-50">
<mml:math id="mml-ieqn-50"><mml:mrow><mml:msub><mml:mi>E</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula> is compared and calculated with <inline-formula id="ieqn-51">
<mml:math id="mml-ieqn-51"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula>, this requires too much calculation. Therefore, MSM adopts an approximate alternative method. Inspired by the candidate selection method [<xref ref-type="bibr" rid="ref-16">16</xref>], MSM samples an entity alignment set <inline-formula id="ieqn-52">
<mml:math id="mml-ieqn-52"><mml:mrow><mml:msub><mml:mi>C</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>c</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>i</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mrow></mml:mrow><mml:mrow><mml:msub><mml:mi>c</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>i</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mrow></mml:mrow><mml:mo>.</mml:mo><mml:mo>.</mml:mo><mml:mo>.</mml:mo><mml:mrow><mml:mo>,</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mi>c</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>i</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mi>c</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>i</mml:mi><mml:mi>t</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msub><mml:mi>E</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:math>
</inline-formula> of <inline-formula id="ieqn-53">
<mml:math id="mml-ieqn-53"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula>, and then calculates the subgraph similarity between <inline-formula id="ieqn-54">
<mml:math id="mml-ieqn-54"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> and these candidate entities. Therefore, for the entity <inline-formula id="ieqn-55">
<mml:math id="mml-ieqn-55"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> in <inline-formula id="ieqn-56">
<mml:math id="mml-ieqn-56"><mml:mrow><mml:msub><mml:mi>E</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula>, the calculation formula for it to be a candidate entity for <inline-formula id="ieqn-57">
<mml:math id="mml-ieqn-57"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> is:</p>
<p><disp-formula id="eqn-5"><label>(5)</label>
<mml:math id="mml-eqn-5" display="block"><mml:mi>p</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mrow><mml:mi>exp</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>&#x22C5;</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msub><mml:mi>E</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mi>exp</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>&#x22C5;</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:math>
</disp-formula>where <inline-formula id="ieqn-58">
<mml:math id="mml-ieqn-58"><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> and <inline-formula id="ieqn-59">
<mml:math id="mml-ieqn-59"><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> are the node vector representations of <inline-formula id="ieqn-60">
<mml:math id="mml-ieqn-60"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> and <inline-formula id="ieqn-61">
<mml:math id="mml-ieqn-61"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> output by MNE.</p>
<p>We propose the following graph matching network, which changes the node update module in each propagation layer. It not only considers the aggregated messages on the edge of each graph as before, but also considers a cross-graph matching vector, which measures the degree of matching between a node in one graph and one or more nodes in another graph. Define <inline-formula id="ieqn-62">
<mml:math id="mml-ieqn-62"><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mrow></mml:mrow><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math>
</inline-formula> as the entity pair to be measured, where <inline-formula id="ieqn-63">
<mml:math id="mml-ieqn-63"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msub><mml:mi>E</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula>, <inline-formula id="ieqn-64">
<mml:math id="mml-ieqn-64"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msub><mml:mi>E</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula>. The cross-matching vector of <inline-formula id="ieqn-65">
<mml:math id="mml-ieqn-65"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> is calculated as follows:</p>
<p><disp-formula id="eqn-6">
<mml:math id="mml-eqn-6" display="block"><mml:mrow><mml:msub><mml:mi>a</mml:mi><mml:mrow><mml:mi>j</mml:mi><mml:mo stretchy="false">&#x2192;</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>=</mml:mo><mml:mi>cos</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>e</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math>
</disp-formula></p>
<p><disp-formula id="eqn-7">
<mml:math id="mml-eqn-7" display="block"><mml:msubsup><mml:mi>h</mml:mi><mml:mi>i</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mrow><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:msubsup><mml:mrow><mml:mrow><mml:msub><mml:mi>a</mml:mi><mml:mrow><mml:mi>j</mml:mi><mml:mo stretchy="false">&#x2192;</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>&#x22C5;</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:mrow><mml:mrow><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:msubsup><mml:mrow><mml:mrow><mml:msub><mml:mi>a</mml:mi><mml:mrow><mml:mi>j</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mo>&gt;</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:math>
</disp-formula></p>
<p><disp-formula id="eqn-8"><label>(6)</label>
<mml:math id="mml-eqn-8" display="block"><mml:mrow><mml:msub><mml:mi>m</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:msubsup><mml:mi>h</mml:mi><mml:mi>i</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math>
</disp-formula></p>
<p>Here <inline-formula id="ieqn-66">
<mml:math id="mml-ieqn-66"><mml:mrow><mml:msub><mml:mi>a</mml:mi><mml:mrow><mml:mi>j</mml:mi><mml:mo stretchy="false">&#x2192;</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:math>
</inline-formula> is the cosine similarities of entity <inline-formula id="ieqn-67">
<mml:math id="mml-ieqn-67"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> in <inline-formula id="ieqn-68">
<mml:math id="mml-ieqn-68"><mml:mrow><mml:msub><mml:mi>E</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula> with the entity <inline-formula id="ieqn-69">
<mml:math id="mml-ieqn-69"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> in <inline-formula id="ieqn-70">
<mml:math id="mml-ieqn-70"><mml:mrow><mml:msub><mml:mi>E</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula>, <inline-formula id="ieqn-71">
<mml:math id="mml-ieqn-71"><mml:mrow><mml:msub><mml:mi>m</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> is the matching vector of node <inline-formula id="ieqn-72">
<mml:math id="mml-ieqn-72"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula>, <inline-formula id="ieqn-73">
<mml:math id="mml-ieqn-73"><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> is the sampled neighbor set of <inline-formula id="ieqn-74">
<mml:math id="mml-ieqn-74"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula>, <inline-formula id="ieqn-75">
<mml:math id="mml-ieqn-75"><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> and <inline-formula id="ieqn-76">
<mml:math id="mml-ieqn-76"><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> are the node vector representations of <inline-formula id="ieqn-77">
<mml:math id="mml-ieqn-77"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> and <inline-formula id="ieqn-78">
<mml:math id="mml-ieqn-78"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> output by MNE, respectively. <inline-formula id="ieqn-79">
<mml:math id="mml-ieqn-79"><mml:msubsup><mml:mi>h</mml:mi><mml:mi>i</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math>
</inline-formula> is an attentive vector for the entire graph <inline-formula id="ieqn-80">
<mml:math id="mml-ieqn-80"><mml:mrow><mml:msub><mml:mi>E</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math>
</inline-formula> . Then, we combine <inline-formula id="ieqn-81">
<mml:math id="mml-ieqn-81"><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> and <inline-formula id="ieqn-82">
<mml:math id="mml-ieqn-82"><mml:mrow><mml:msub><mml:mi>m</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula> to get the sampled neighbor representation of <inline-formula id="ieqn-83">
<mml:math id="mml-ieqn-83"><mml:mrow><mml:msub><mml:mi>e</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</inline-formula>:</p>
<p><disp-formula id="eqn-9"><label>(7)</label>
<mml:math id="mml-eqn-9" display='block'><mml:mrow><mml:msub><mml:mover accent='true'><mml:mi>h</mml:mi><mml:mo>&#x2322;</mml:mo></mml:mover><mml:mi>i</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo>&#x2295;</mml:mo><mml:mi>&#x03B2;</mml:mi><mml:mo>*</mml:mo><mml:msub><mml:mi>m</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:math>
</disp-formula>where <inline-formula id="ieqn-84">
<mml:math id="mml-ieqn-84"><mml:mo>&#x2295;</mml:mo></mml:math>
</inline-formula> indicates vector concatenation.</p>
</sec>
<sec id="s2_4">
<label>2.4</label>
<title>Entity Alignment</title>
<p>We give the loss function of MNE:</p>
<p><disp-formula id="eqn-10"><label>(8)</label>
<mml:math id="mml-eqn-10" display="block"><mml:mi>L</mml:mi><mml:mo>=</mml:mo><mml:munder><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:mi mathvariant="normal">X</mml:mi></mml:mrow></mml:mrow></mml:munder><mml:mrow><mml:munder><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mi>i</mml:mi><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:msup><mml:mi>j</mml:mi><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi mathvariant="normal">X</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mrow></mml:munder><mml:mrow><mml:mo movablelimits="true" form="prefix">max</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mrow></mml:mrow><mml:mi>d</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mi>d</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mi>i</mml:mi><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msup><mml:mi>j</mml:mi><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo>+</mml:mo><mml:mi>&#x03B3;</mml:mi></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mrow></mml:mrow></mml:math>
</disp-formula>where <inline-formula id="ieqn-85">
<mml:math id="mml-ieqn-85"><mml:mi>d</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msub><mml:mi>h</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:mrow></mml:math>
</inline-formula>, <inline-formula id="ieqn-86">
<mml:math id="mml-ieqn-86"><mml:mrow><mml:msub><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:mrow><mml:mrow></mml:mrow><mml:mo>&#x22C5;</mml:mo><mml:mrow></mml:mrow></mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:mrow></mml:math>
</inline-formula> means <inline-formula id="ieqn-87">
<mml:math id="mml-ieqn-87"><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mi>n</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>m</mml:mi></mml:math>
</inline-formula>; <inline-formula id="ieqn-88">
<mml:math id="mml-ieqn-88"><mml:mi>&#x03B3;</mml:mi><mml:mo>&gt;</mml:mo><mml:mn>0</mml:mn></mml:math>
</inline-formula> is a margin hyper-parameter; <inline-formula id="ieqn-89">
<mml:math id="mml-ieqn-89"><mml:mrow><mml:mi mathvariant="normal">X</mml:mi></mml:mrow></mml:math>
</inline-formula> is the alignment seeds and <inline-formula id="ieqn-90">
<mml:math id="mml-ieqn-90"><mml:mrow><mml:msup><mml:mrow><mml:mi mathvariant="normal">X</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:math>
</inline-formula> is the set of negative aligned entity pairs generated by nearest neighbor sampling [<xref ref-type="bibr" rid="ref-15">15</xref>].</p>
<p>The loss function of MSM after the pre-training phase, defined as:</p>
<p><disp-formula id="eqn-11"><label>(9)</label>
<mml:math id="mml-eqn-11" display="block"><mml:mi>L</mml:mi><mml:mo>=</mml:mo><mml:munder><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2208;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:munder><mml:mrow><mml:munder><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mi>i</mml:mi><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:msup><mml:mi>j</mml:mi><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msup><mml:mi>T</mml:mi><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mrow></mml:munder><mml:mrow><mml:mo movablelimits="true" form="prefix">max</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mtext>&#xA0;</mml:mtext><mml:mi>f</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mi>f</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mi>i</mml:mi><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:msup><mml:mi>j</mml:mi><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo>+</mml:mo><mml:mi>&#x03B3;</mml:mi></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mrow></mml:mrow></mml:math>
</disp-formula></p>
<p>where <inline-formula id="ieqn-85a"><mml:math id="mml-ieqn-85a"><mml:mrow><mml:mi>f</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mrow><mml:mo>&#x2016;</mml:mo><mml:mrow><mml:msub><mml:mover accent='true'><mml:mi>h</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mrow><mml:mrow><mml:mrow><mml:msub><mml:mover accent='true'><mml:mi>h</mml:mi><mml:mo>&#x2322;</mml:mo></mml:mover><mml:mi>j</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2016;</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula>; <inline-formula id="ieqn-91">
<mml:math id="mml-ieqn-91"><mml:mi>T</mml:mi></mml:math>
</inline-formula> is the positive alignments set and <inline-formula id="ieqn-92">
<mml:math id="mml-ieqn-92"><mml:mrow><mml:msup><mml:mi>T</mml:mi><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:math>
</inline-formula> the negative alignments set.</p>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Experiment</title>
<p>In this section, we evaluated the multi-neighborhood sampling matching model (MSM) on three data subsets of the open-source data set DBP15k by comparing with other methods. <xref ref-type="table" rid="table-1">Tab. 1</xref> gives detailed descriptions of the DBP15k datasets. We use the same split with previous work, 30% for training and 70% for testing.</p>
<table-wrap id="table-1"><label>Table 1</label>
<caption>
<title>Summary of DBP15K</title></caption>
<table><colgroup>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Datasets</th>
<th>Ent</th>
<th>Rel</th>
<th>Tri</th>
</tr>
</thead>
<tbody>
<tr>
<td>ZH-EN ZH</td>
<td>66,469</td>
<td>2,830</td>
<td>153,929</td>
</tr>
<tr>
<td>EN</td>
<td>98,125</td>
<td>2,317</td>
<td>237,674</td>
</tr>
<tr>
<td>JA-EN JA</td>
<td>65,744</td>
<td>2,043</td>
<td>164,373</td>
</tr>
<tr>
<td>EN</td>
<td>95,680</td>
<td>2,096</td>
<td>233,319</td>
</tr>
<tr>
<td>FR-EN FR</td>
<td>66,858</td>
<td>1,379</td>
<td>192,191</td>
</tr>
<tr>
<td>EN</td>
<td>105,889</td>
<td>2,209</td>
<td>278,590</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The configuration we use in the DBP15K datasets is: <inline-formula id="ieqn-93">
<mml:math id="mml-ieqn-93"><mml:mi>&#x03B2;</mml:mi><mml:mo>=</mml:mo><mml:mn>0.1</mml:mn></mml:math>
</inline-formula>, <inline-formula id="ieqn-94">
<mml:math id="mml-ieqn-94"><mml:mi>&#x03B3;</mml:mi><mml:mo>=</mml:mo><mml:mn>1.0</mml:mn></mml:math>
</inline-formula>, and we sample 5 neighbors for each entity in the neighborhood sampling stage.</p>
<p>The comparison method we use is as follows:</p>
<p>(1) JAPE [<xref ref-type="bibr" rid="ref-17">17</xref>]: A model for preserving attribute embedding for cross-language entity alignment. It embeds the structure of the two networks into a unified vector space, and further optimizes it by using the attribute correlation in the network. This model is a supervised learning model. In the experiment, 30% of the nodes in the data set are used as training data and 70% of the nodes are used as test data. The model parameters adopt the original default best parameter combination. In the experiment, 10 negative samples are drawn for each pre-aligned entity pair.</p>
<p>(2) AliNet [<xref ref-type="bibr" rid="ref-18">18</xref>]: In order to solve a new KG alignment network in which the replica entity has a non-isomorphic neighbor structure, it aims to alleviate the non-isomorphism problem of the neighbor structure in an end-to-end manner. Due to the heterogeneity of the schema, the direct neighbors of the replica entity are usually not similar. AliNet introduces distant neighbors to extend the overlapping part of the neighbor structure. The attention mechanism is also used to emphasize helpful distant neighbors and reduce noise. Then it uses the gate mechanism to aggregate the information of direct neighbors and distant neighbors.</p>
<p>(3) BoostEA [<xref ref-type="bibr" rid="ref-16">16</xref>]: The entity alignment based on the embedded usually depends on the alignment of the existing entities as training data. But the first alignment can be usually accounted for a small part. BoostEA proposes a bootstrapping method to solve the above challenges.</p>
<p>(4) GMNN [<xref ref-type="bibr" rid="ref-5">5</xref>]: Since the previous cross-language knowledge graph alignment research relies on the idea of entity embedding, it cannot be applied to the two knowledge graphs. GMNN combines the two kinds of map matching problems with entity context information into a map matching problem and proposes a matching model of a graph neural network, which includes map matching and pixel matching information.</p>
<p>(5) MSM: Firstly, the low-dimensional vector representation of nodes in the network is obtained by using the multi-neighborhood network representation learning method (MNE) proposed in Section 2 of this paper. Then, entity alignment is carried out by using the information entropy sampling and cross-graph neighborhood matching entity alignment method proposed in this paper.</p>
<sec id="s3_1">
<label>3.1</label>
<title>Evaluation Index</title>
<p>To compares the various methods more objectively and comprehensively, in the experiment, the commonly used model evaluation indicators <inline-formula id="ieqn-95">
<mml:math id="mml-ieqn-95"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mrow><mml:mo>@</mml:mo></mml:mrow><mml:mi>k</mml:mi></mml:math>
</inline-formula> in the field of entity alignment are used, and the calculation formula is as follows:</p>
<p><disp-formula id="eqn-12"><label>(11)</label>
<mml:math id="mml-eqn-12" display="block"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>@</mml:mo></mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mrow><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>r</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow><mml:mi>N</mml:mi></mml:mfrac></mml:mrow></mml:mstyle></mml:math>
</disp-formula>where <inline-formula id="ieqn-96">
<mml:math id="mml-ieqn-96"><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>r</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:math>
</inline-formula> represents the number of times the target node appears in the <inline-formula id="ieqn-97">
<mml:math id="mml-ieqn-97"><mml:mi>k</mml:mi></mml:math>
</inline-formula> closest candidate target nodes, and <inline-formula id="ieqn-98">
<mml:math id="mml-ieqn-98"><mml:mi>N</mml:mi></mml:math>
</inline-formula> is the number of tests. We select <inline-formula id="ieqn-99">
<mml:math id="mml-ieqn-99"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>@</mml:mo></mml:mrow><mml:mn>1</mml:mn></mml:math>
</inline-formula> and <inline-formula id="ieqn-100">
<mml:math id="mml-ieqn-100"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>@</mml:mo></mml:mrow><mml:mn>10</mml:mn></mml:math>
</inline-formula> indicators according to the commonly used experimental methods.</p>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Analysis of Results</title>
<p>(1) The experimental results of each method on different data subsets of the data set DBP15k are shown in <xref ref-type="table" rid="table-2">Tab. 2</xref>. From <xref ref-type="table" rid="table-2">Tab. 2</xref>, we can see that under the evaluation index of <inline-formula id="ieqn-101">
<mml:math id="mml-ieqn-101"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>@</mml:mo></mml:mrow><mml:mn>1</mml:mn></mml:math>
</inline-formula>, MSM has an improvement of 3% to 5% relative to GMNN, and also has an improvement of about 3% under the evaluation index of <inline-formula id="ieqn-102">
<mml:math id="mml-ieqn-102"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>@</mml:mo></mml:mrow><mml:mn>10</mml:mn></mml:math>
</inline-formula>. The result meets 8% improvement on zh-en data subset. Compared with JAPE, AliNet and BoostEA, MSM has a greater improvement.</p>
<p>To evaluate the sampling method on the accuracy of the model, we implement a model variant (referred as MSM <sub>(s-two)</sub>) that samples two-hop neighbors. From <xref ref-type="table" rid="table-2">Tab. 2</xref>, we find that the sampling two-hop neighbors leads MSM a gain of 0.4%&#x007E;0.8% by <inline-formula id="ieqn-103">
<mml:math id="mml-ieqn-103"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>@</mml:mo></mml:mrow><mml:mn>1</mml:mn></mml:math>
</inline-formula> and 0.1%&#x007E;0.3% by <inline-formula id="ieqn-104">
<mml:math id="mml-ieqn-104"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>@</mml:mo></mml:mrow><mml:mn>10</mml:mn></mml:math>
</inline-formula> relative to the sampling one-hop neighbors (referred as MSM <sub>(s-one)</sub>) on DBP15K. The result confirms the importance of the two-hop neighborhoods in KG structure embedding and neighborhood sampling.</p>
<table-wrap id="table-2"><label>Table 2</label>
<caption>
<title>Performance on DBP15K (%)</title></caption>
<table><colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th rowspan="2">Models</th>
<th colspan="2">DBPzh-en</th>
<th colspan="2">DBPja-en</th>
<th colspan="2">DBPfr-en</th>
</tr>
<tr>
<th>Hits@1</th>
<th>Hits@10</th>
<th>Hits@1</th>
<th>Hits@10</th>
<th>Hits@1</th>
<th>Hits@10</th>
</tr>
</thead>
<tbody>
<tr>
<td>JAPE</td>
<td>41.2</td>
<td>74.5</td>
<td>36.3</td>
<td>68.5</td>
<td>32.3</td>
<td>66.7</td>
</tr>
<tr>
<td>AliNet</td>
<td>53.9</td>
<td>82.6</td>
<td>54.9</td>
<td>83.1</td>
<td>55.2</td>
<td>85.2</td>
</tr>
<tr>
<td>BootEA</td>
<td>62.9</td>
<td>84.8</td>
<td>62.2</td>
<td>85.4</td>
<td>65.3</td>
<td>87.4</td>
</tr>
<tr>
<td>GMNN</td>
<td>67.9</td>
<td>78.5</td>
<td>74.0</td>
<td>87.2</td>
<td>89.4</td>
<td>95.2</td>
</tr>
<tr>
<td>MSM<sub>(s-one)</sub></td>
<td>70.7</td>
<td><bold>86.9</bold></td>
<td>78.2</td>
<td>91.4</td>
<td><bold>90.6</bold></td>
<td>97.1</td>
</tr>
<tr>
<td>MSM<sub>(s-two)</sub></td>
<td><bold>70.8</bold></td>
<td>86.7</td>
<td><bold>78.8 </bold></td>
<td><bold>91.9</bold></td>
<td><bold>90.6 </bold></td>
<td><bold>97.3</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>(2) To explore the impact of using the pre-trained word embeddings to initialize the MSM. We remove the initialization part of the MNE, choosing the different sampling size for comparative analysis. From <xref ref-type="fig" rid="fig-3">Fig. 3</xref>, for DBP15K<sub>ZH-EN</sub>, we observe that the initialization of MNE is very important of MSM. More exactly, removing the initialization from MSM, leads to around a 43% drop in <inline-formula id="ieqn-105">
<mml:math id="mml-ieqn-105"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>@</mml:mo></mml:mrow><mml:mn>1</mml:mn></mml:math>
</inline-formula> and a 40% drop in<inline-formula id="ieqn-106">
<mml:math id="mml-ieqn-106"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>@</mml:mo></mml:mrow><mml:mn>10</mml:mn></mml:math>
</inline-formula> on average. On the other hand, MSM is sensitive to the size of sampling and the model works better when the size of sampling is 5.</p>
<fig id="fig-3">
<label>Figure 3</label>
<caption>
<title>Comparison the pre-training with removing pre-training. a) The evaluation index of <inline-formula id="ieqn-107">
<mml:math id="mml-ieqn-107"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>@</mml:mo></mml:mrow><mml:mn>1</mml:mn></mml:math>
</inline-formula> b) The evaluation index of <inline-formula id="ieqn-108">
<mml:math id="mml-ieqn-108"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>@</mml:mo></mml:mrow><mml:mn>10</mml:mn></mml:math>
</inline-formula></title></caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_20218-fig-3.png"/>
</fig>
<p>(3) To verify the superiority of our cross-graph neighborhood matching method, we compare it with the common direct aggregate matching method. As shown in <xref ref-type="fig" rid="fig-4">Fig. 4</xref>, when we choose different candidate set sizes, the cosine cross-neighborhood matching method performs better. It increases 0.5%&#x007E;1% relative to the overall the cross-neighborhood matching method under the evaluation index of <inline-formula id="ieqn-109">
<mml:math id="mml-ieqn-109"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>@</mml:mo></mml:mrow><mml:mn>10</mml:mn></mml:math>
</inline-formula> and about 0.2% under the evaluation index of <inline-formula id="ieqn-110">
<mml:math id="mml-ieqn-110"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>@</mml:mo></mml:mrow><mml:mn>1</mml:mn></mml:math>
</inline-formula>. At the same time, the different matching candidate size also has some influence on MSM performance. When the size of candidate selection is 18, the performance of the MSM is better on DBP15K<sub>ZH-EN</sub>.</p>
<fig id="fig-4">
<label>Figure 4</label>
<caption>
<title>Comparison different matching methods. a) The evaluation index of <inline-formula id="ieqn-111">
<mml:math id="mml-ieqn-111"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>@</mml:mo></mml:mrow><mml:mn>1</mml:mn></mml:math>
</inline-formula> b) The evaluation index of <inline-formula id="ieqn-112">
<mml:math id="mml-ieqn-112"><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi><mml:mi>s</mml:mi><mml:mrow><mml:mo>@</mml:mo></mml:mrow><mml:mn>10</mml:mn></mml:math>
</inline-formula></title></caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_20218-fig-4.png"/>
</fig>
<p>(4) To show the effect of entity alignment more intuitively, this part visually shows the node distribution consistency before and after entity alignment. We reduce the 300-dimensional vector to two-dimensional space through t-sne. <xref ref-type="fig" rid="fig-5">Fig. 5</xref> show the visualization of 4000 pairs of nodes randomly selected from two KG on DBP15K<sub>JA-EN</sub>.</p>
<fig id="fig-5">
<label>Figure 5</label>
<caption>
<title>Visualization of spatial distribution of entity alignment nodes in DBP15K<sub>JA-EN</sub> dataset. a) Initial state b) After 200 rounds of training c) After 400 rounds of training</title></caption>
<graphic mimetype="image" mime-subtype="png" xlink:href="IASC_20218-fig-5.png"/>
</fig>
<p>These result show that our pre-trained vectors, sampling and matching modules are particularly important, when the neighborhood sizes of equivalent entities greatly differ and especially there may be few common neighbors in their neighborhoods.</p>
</sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Conclusion</title>
<p>This paper proposes a multi-neighborhood sampling matching entity alignment method, which aims to solve the problem of different neighborhood sizes and topological structures in heterogeneous networks. We build a multi-neighborhood network representation learning method to achieve effective aggregation of entity neighborhood information and use a new sampling-based method to select the most informative neighbor for each entity, using the method of cosine cross-graph neighborhood matching to achieve rapid alignment of different network entities. We conducted extensive experiments on real-world data sets and compared MSM with four embedded-based entity alignment methods. Experimental results show that MSM obtains the best and more robust performance, and consistently outperforms competing methods in data sets and evaluation metrics. For future work, we plan to incorporate the multi-neighbor information of entities in other modes into our model structure. At the same time, since some alterative sampling techniques based on ranks information may lead more efficient procedure [<xref ref-type="bibr" rid="ref-19">19</xref>,<xref ref-type="bibr" rid="ref-20">20</xref>], this will be our follow-up work.</p>
</sec>
</body>
<back>
<ack>
<p>The authors are grateful to the anonymous referees for having carefully read earlier versions of the manuscript. Their valuable suggestions substantially improved the quality of exposition, shape, and content of the article.</p>
</ack><fn-group>
<fn fn-type="other">
<p><bold>Funding Statement:</bold> This work is supported by State Grid Shandong Electric Power Company Science and Technology Project Funding under Grant Nos. 520613200001, 520613180002, 62061318C002, Weihai Scientific Research and Innovation Fund (2020) and the Grant 19YG02, Sanming University.</p>
</fn>
<fn fn-type="conflict">
<p><bold>Conflicts of Interest:</bold> The authors declare that they have no conflicts of interest to report regarding the present study.</p>
</fn>
</fn-group>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1"><label>1</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>D.</given-names> <surname>Zhu</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Wang</surname></string-name> and <string-name><given-names>C.</given-names> <surname>You</surname></string-name></person-group>, &#x201C;<article-title>MMLUP: Multi-source &#x0026; multi-task learning for user profiles in social network</article-title>,&#x201D; <source>Computers, Materials &#x0026; Continua</source>, vol. <volume>61</volume>, no. <issue>3</issue>, pp. <fpage>1105</fpage>&#x2013;<lpage>1115</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-2"><label>2</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>You</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Zhu</surname></string-name> and <string-name><given-names>Y.</given-names> <surname>Sun</surname></string-name></person-group>, &#x201C;<article-title>SNES: Social-network-oriented public opinion monitoring platform based on ElasticSearch</article-title>,&#x201D; <source>Computers, Materials &#x0026; Continua</source>, vol. <volume>61</volume>, no. <issue>3</issue>, pp. <fpage>1271</fpage>&#x2013;<lpage>1283</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-3"><label>3</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>N.</given-names> <surname>Jia</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Tian</surname></string-name> and <string-name><given-names>Y. Y.</given-names> <surname>Zhang</surname></string-name></person-group>, &#x201C;<article-title>Semi-supervised node classification with discriminable squeeze excitation graph convolutional networks</article-title>,&#x201D; <source>IEEE Access</source>, vol. <volume>8</volume>, pp. <fpage>148226</fpage>&#x2013;<lpage>148236</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-4"><label>4</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>D.</given-names> <surname>Zhu</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Sun</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Du</surname></string-name>, <string-name><given-names>N.</given-names> <surname>Cao</surname></string-name>, <string-name><given-names>T.</given-names> <surname>Baker</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>HUNA: A method of hierarchical unsupervised network alignment for IoT</article-title>,&#x201D; <source>IEEE Internet of Things Journal</source>, vol. <volume>8</volume>, no. <issue>5</issue>, pp. <fpage>3201</fpage>&#x2013;<lpage>3210</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-5"><label>5</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>K.</given-names> <surname>Xu</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Wang</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Yu</surname></string-name></person-group>, &#x201C;<article-title>Cross-lingual knowledge graph alignment via graph matching neural network</article-title>,&#x201D; in <conf-name>57th Annual Meeting of the Association-for-Computational-Linguistics</conf-name>, <publisher-loc>Florence, ITALY</publisher-loc>, pp. <fpage>3156</fpage>&#x2013;<lpage>3161</lpage>, <year>2019</year>. </mixed-citation></ref>
<ref id="ref-6"><label>6</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Zhu</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Xie</surname></string-name> and <string-name><given-names>Z.</given-names> <surname>Liu</surname></string-name></person-group>, &#x201C;<article-title>Iterative entity alignment via joint knowledge embeddings</article-title>,&#x201D; in <conf-name>26th Int. Joint Conf. on Artificial Intelligence</conf-name>, <publisher-loc>Melbourne, Australia</publisher-loc>, vol. <volume>17</volume>, pp. <fpage>4258</fpage>&#x2013;<lpage>4264</lpage>, <year>2017</year>. </mixed-citation></ref>
<ref id="ref-7"><label>7</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>D.</given-names> <surname>Zhu</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Sun</surname></string-name> and <string-name><given-names>X.</given-names> <surname>Li</surname></string-name></person-group>, &#x201C;<article-title>MINE: A method of multi-Interaction heterogeneous information network embedding</article-title>,&#x201D; <source>Computers, Materials &#x0026; Continua</source>, vol. <volume>63</volume>, no. <issue>3</issue>, pp. <fpage>1343</fpage>&#x2013;<lpage>1356</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-8"><label>8</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Cao</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Liu</surname></string-name> and <string-name><given-names>C.</given-names> <surname>Li</surname></string-name></person-group>, &#x201C;<article-title>Multi-channel graph neural network for entity alignment</article-title>,&#x201D; in <conf-name>57th Annual Meeting of the Association-for-Computational-Linguistics</conf-name>, <publisher-loc>Florence, Italy</publisher-loc>, pp. <fpage>1452</fpage>&#x2013;<lpage>1461</lpage>, <year>2019</year>. </mixed-citation></ref>
<ref id="ref-9"><label>9</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Q.</given-names> <surname>Zhu</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Zhou</surname></string-name> and <string-name><given-names>J.</given-names> <surname>Wu</surname></string-name></person-group>, &#x201C;<article-title>Neighborhood-aware attentional representation for multilingual knowledge graphs</article-title>,&#x201D; in <conf-name>28th Int. Joint Conf. on Artificial Intelligence</conf-name>, <publisher-loc>Macau, China</publisher-loc>, pp. <fpage>1943</fpage>&#x2013;<lpage>1949</lpage>, <year>2019</year>. </mixed-citation></ref>
<ref id="ref-10"><label>10</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Guan</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Jin</surname></string-name> and <string-name><given-names>Y.</given-names> <surname>Wang</surname></string-name></person-group>, &#x201C;<article-title>Self-learning and embedding based entity alignment</article-title>,&#x201D; <source>Knowledge and Information Systems</source>, vol. <volume>59</volume>, no. <issue>2</issue>, pp. <fpage>316</fpage>&#x2013;<lpage>386</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-11"><label>11</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Yan</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Peng</surname></string-name> and <string-name><given-names>Y.</given-names> <surname>Wang</surname></string-name></person-group>, &#x201C;<article-title>CTEA: Context and topic enhanced entity alignment for knowledge graphs</article-title>,&#x201D; <source>Neurocomputing</source>, vol. <volume>410</volume>, pp. <fpage>419</fpage>&#x2013;<lpage>431</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-12"><label>12</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Li</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Cao</surname></string-name> and <string-name><given-names>L.</given-names> <surname>Hou</surname></string-name></person-group>, &#x201C;<article-title>Semi-supervised entity alignment via joint knowledge embedding model and cross-graph model</article-title>,&#x201D; in <conf-name>2019 Conf. on Empirical Methods in Natural Language Processing and 9th Int. Joint Conf. on Natural Language Processing</conf-name>, <conf-loc>Hong Kong, China</conf-loc>, vol. <volume>1</volume>, pp. <fpage>2723</fpage>&#x2013;<lpage>2732</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-13"><label>13</label><mixed-citation publication-type="other"><person-group person-group-type="author">Y. Yan, L. Liu and Y. Ban</person-group>, &#x201C;<article-title>Dynamic Knowledge Graph Alignment</article-title>,&#x201D; in <italic>Proceedings of the AAAI Conference on Artificial Intelligence</italic>, vol. <volume>35</volume>, no. <issue>5</issue>, pp. <fpage>4564</fpage>&#x2013;<lpage>4572</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-14"><label>14</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Pei</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Yu</surname></string-name> and <string-name><given-names>X.</given-names> <surname>Zhang</surname></string-name></person-group>, &#x201C;<article-title>Improving cross-lingual entity alignment via optimal transport</article-title>,&#x201D; in <conf-name>Int. Joint Conf. on Artificial Intelligence Organization</conf-name>, <conf-loc>Macao, China</conf-loc>, vol. <volume>1</volume>, pp. <fpage>3231</fpage>&#x2013;<lpage>3237</lpage>, <year>2019</year>. </mixed-citation></ref>
<ref id="ref-15"><label>15</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Wu</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Liu</surname></string-name> and <string-name><given-names>Y.</given-names> <surname>Feng</surname></string-name></person-group>, &#x201C;<article-title>Neighborhood matching network for entity alignment</article-title>,&#x201D; in <conf-name>58th Annual Meeting of the Association for Computational Linguistics</conf-name>, <publisher-loc>Washington, USA</publisher-loc>, vol. <volume>1</volume>, pp. <fpage>6477</fpage>&#x2013;<lpage>6487</lpage>, <year>2020</year>. </mixed-citation></ref>
<ref id="ref-16"><label>16</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Sun</surname></string-name>, <string-name><given-names>W.</given-names> <surname>Hu</surname></string-name>, <string-name><given-names>Q.</given-names> <surname>Zhang</surname></string-name> and <string-name><given-names>Y.</given-names> <surname>Qu</surname></string-name></person-group>, &#x201C;<article-title>Bootstrapping entity alignment with knowledge graph embedding</article-title>,&#x201D; in <conf-name>27th Int. Joint Conf. on Artificial Intelligence</conf-name>, <publisher-loc>Macau, China</publisher-loc>, vol. <volume>18</volume>, pp. <fpage>4396</fpage>&#x2013;<lpage>4402</lpage>, <year>2018</year>. </mixed-citation></ref>
<ref id="ref-17"><label>17</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Sun</surname></string-name>, <string-name><given-names>W.</given-names> <surname>Hu</surname></string-name> and <string-name><given-names>C.</given-names> <surname>Li</surname></string-name></person-group>, &#x201C;<article-title>Cross-lingual entity alignment via joint attribute-preserving embedding</article-title>,&#x201D; in <conf-name>Int. Semantic Web Conf.</conf-name>, <publisher-loc>Cham</publisher-loc>, <publisher-name>Springer</publisher-name>, <year>2017</year>. </mixed-citation></ref>
<ref id="ref-18"><label>18</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Sun</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Wang</surname></string-name> and <string-name><given-names>W.</given-names> <surname>Hu</surname></string-name></person-group>, &#x201C;<article-title>Knowledge graph alignment network with gated multi-hop neighborhood aggregation</article-title>,&#x201D; in <conf-name>34th AAAI Conf. on Artificial Intelligence</conf-name>, <publisher-loc>New York, USA</publisher-loc>, vol. <volume>34</volume>, pp. <fpage>222</fpage>&#x2013;<lpage>229</lpage>, <year>2020</year>. </mixed-citation></ref>
<ref id="ref-19"><label>19</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Mahdizadeh</surname></string-name> and <string-name><given-names>E.</given-names> <surname>Zamanzade</surname></string-name></person-group>, &#x201C;<article-title>Kernel-based estimation of P (X &#x003E; Y) in ranked set sampling</article-title>,&#x201D; <source>SORT-Statistics and Operations Research Transactions</source>, vol. <volume>1</volume>, no. <issue>2</issue>, pp. <fpage>243</fpage>&#x2013;<lpage>266</lpage>, <year>2016</year>.</mixed-citation></ref>
<ref id="ref-20"><label>20</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Mahdizadeh</surname></string-name> and <string-name><given-names>E.</given-names> <surname>Zamanzade</surname></string-name></person-group>, &#x201C;<article-title>Smooth estimation of a reliability function in ranked set sampling</article-title>,&#x201D; <source>Statistics</source>, vol. <volume>52</volume>, no. <issue>4</issue>, pp. <fpage>750</fpage>&#x2013;<lpage>768</lpage>, <year>2018</year>.</mixed-citation></ref>
</ref-list>
</back>
</article>