<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Blog | Enjun Du（杜恩俊）</title><link>https://enjundu.com/blog/</link><atom:link href="https://enjundu.com/blog/index.xml" rel="self" type="application/rss+xml"/><description>Blog</description><generator>Hugo Blox Builder (https://hugoblox.com)</generator><language>en-us</language><lastBuildDate>Tue, 25 Mar 2025 00:00:00 +0000</lastBuildDate><image><url>https://enjundu.com/media/icon_hu7729264130191091259.png</url><title>Blog</title><link>https://enjundu.com/blog/</link></image><item><title>GraphOracle: Efficient Fully-Inductive Knowledge Graph Reasoning via Relation-Dependency Graphs</title><link>https://enjundu.com/blog/graphoracle/</link><pubDate>Tue, 25 Mar 2025 00:00:00 +0000</pubDate><guid>https://enjundu.com/blog/graphoracle/</guid><description>&lt;h2 id="abstract">Abstract&lt;/h2>
&lt;p>Knowledge graph reasoning in the &lt;strong>fully-inductive setting&lt;/strong> — where both entities and relations at test time are unseen during training — remains an open challenge. We introduce &lt;strong>GraphOracle&lt;/strong>, a novel framework that transforms each knowledge graph into a &lt;strong>Relation-Dependency Graph (RDG)&lt;/strong> encoding directed precedence links between relations. A &lt;strong>multi-head attention&lt;/strong> mechanism produces context-aware relation embeddings that guide inductive message passing over the original KG. Experiments on &lt;strong>60 benchmarks&lt;/strong> show &lt;strong>up to 25% improvement&lt;/strong> in fully-inductive and &lt;strong>28% in cross-domain&lt;/strong> scenarios.&lt;/p>
&lt;hr>
&lt;h2 id="motivation">Motivation&lt;/h2>
&lt;p>Existing fully-inductive methods (INGRAM, ULTRA) construct &lt;strong>undirected&lt;/strong> relation graphs that are dense and fail to capture &lt;strong>directed compositional patterns&lt;/strong> — e.g., &amp;ldquo;born_in → located_in&amp;rdquo; is a directional dependency that undirected graphs cannot distinguish. Furthermore, they are limited to single-domain scenarios and cannot transfer across entirely different knowledge graphs.&lt;/p>
&lt;hr>
&lt;h2 id="method">Method&lt;/h2>
&lt;p>
&lt;figure >
&lt;div class="flex justify-center ">
&lt;div class="w-100" >&lt;img alt="Framework" srcset="
/blog/graphoracle/framework_hu17442796894096862171.webp 400w,
/blog/graphoracle/framework_hu16880348516718552876.webp 760w,
/blog/graphoracle/framework_hu17354898415690024261.webp 1200w"
src="https://enjundu.com/blog/graphoracle/framework_hu17442796894096862171.webp"
width="760"
height="303"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>GraphOracle operates in three stages:&lt;/p>
&lt;ul>
&lt;li>&lt;strong>RDG Construction&lt;/strong> — Transform the KG into a directed Relation-Dependency Graph where edge &lt;em>(r_i, r_j)&lt;/em> indicates that relation &lt;em>r_i&lt;/em> precedes &lt;em>r_j&lt;/em> in observed triple chains. This is significantly sparser than prior relation graphs while capturing compositional patterns.&lt;/li>
&lt;li>&lt;strong>Query-Dependent Multi-Head Attention&lt;/strong> — For each query relation, a multi-head attention GNN propagates messages over the RDG to produce context-aware relation embeddings. The same relation gets different representations depending on the query context.&lt;/li>
&lt;li>&lt;strong>Entity-Level Answer Prediction&lt;/strong> — The learned relation embeddings parameterize a second GNN on the original KG entity graph, performing inductive message passing from the query entity to score candidates.&lt;/li>
&lt;/ul>
&lt;hr>
&lt;h2 id="experimental-results">Experimental Results&lt;/h2>
&lt;p>Evaluated across &lt;strong>60 benchmarks&lt;/strong> spanning transductive, entity-inductive, fully-inductive, and cross-domain settings.&lt;/p>
&lt;h3 id="mrr-comparison-on-60-datasets">MRR Comparison on 60 Datasets&lt;/h3>
&lt;p>
&lt;figure >
&lt;div class="flex justify-center ">
&lt;div class="w-100" >&lt;img alt="MRR Comparison" srcset="
/blog/graphoracle/mrr_comparison_hu8110470002604910957.webp 400w,
/blog/graphoracle/mrr_comparison_hu4321970991457367304.webp 760w,
/blog/graphoracle/mrr_comparison_hu128044742679316233.webp 1200w"
src="https://enjundu.com/blog/graphoracle/mrr_comparison_hu8110470002604910957.webp"
width="760"
height="272"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>GraphOracle (red) consistently matches or outperforms supervised SOTA (green) across all 60 datasets, with particularly strong gains on cross-domain and biomedical KGs.&lt;/p>
&lt;h3 id="average-performance-4-settings">Average Performance (4 Settings)&lt;/h3>
&lt;p>
&lt;figure >
&lt;div class="flex justify-center ">
&lt;div class="w-100" >&lt;img alt="Main Results" srcset="
/blog/graphoracle/main_results_hu16923454802355985960.webp 400w,
/blog/graphoracle/main_results_hu7398350365299240702.webp 760w,
/blog/graphoracle/main_results_hu2897405029583734145.webp 1200w"
src="https://enjundu.com/blog/graphoracle/main_results_hu16923454802355985960.webp"
width="760"
height="377"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>GraphOracle achieves &lt;strong>+7.19%&lt;/strong> MRR improvement in transductive, &lt;strong>+10.86%&lt;/strong> in entity-inductive, &lt;strong>+13.28%&lt;/strong> in fully-inductive, and &lt;strong>+26.82%&lt;/strong> in cross-domain settings over supervised SOTA.&lt;/p>
&lt;h3 id="ablation-study">Ablation Study&lt;/h3>
&lt;p>
&lt;figure >
&lt;div class="flex justify-center ">
&lt;div class="w-100" >&lt;img alt="Ablation" srcset="
/blog/graphoracle/ablation_hu6660703363372269825.webp 400w,
/blog/graphoracle/ablation_hu3398867998203931436.webp 760w,
/blog/graphoracle/ablation_hu14131753493942245603.webp 1200w"
src="https://enjundu.com/blog/graphoracle/ablation_hu6660703363372269825.webp"
width="760"
height="220"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>Removing the RDG structure or multi-head attention causes significant performance drops, confirming both components are essential. The directed precedence encoding is the single most important design choice.&lt;/p></description></item><item><title>GraphMaster: Automated Graph Synthesis via LLM Agents in Data-Limited Environments</title><link>https://enjundu.com/blog/graphmaster/</link><pubDate>Thu, 06 Mar 2025 00:00:00 +0000</pubDate><guid>https://enjundu.com/blog/graphmaster/</guid><description>&lt;h2 id="abstract">Abstract&lt;/h2>
&lt;p>The era of foundation models has revolutionized AI research, yet &lt;strong>Graph Foundation Models (GFMs)&lt;/strong> remain constrained by the scarcity of large-scale graph corpora. We introduce &lt;strong>GraphMaster&lt;/strong> — the first multi-agent framework specifically designed for graph data synthesis in data-limited environments. GraphMaster orchestrates &lt;strong>four specialized LLM agents&lt;/strong> (Manager, Perception, Enhancement, and Evaluation) that collaboratively optimize the synthesis process through iterative refinement, ensuring both semantic coherence and structural integrity.&lt;/p>
&lt;hr>
&lt;h2 id="motivation">Motivation&lt;/h2>
&lt;p>Graph Foundation Models face a critical data bottleneck. Existing synthesis methods fail for three reasons: &lt;strong>(1)&lt;/strong> classical augmentation (GraphSmote, G-Mixup) only manipulates structure, producing semantically empty nodes; &lt;strong>(2)&lt;/strong> LLMs cannot process entire graphs within context windows; &lt;strong>(3)&lt;/strong> uncoordinated LLM generation introduces hallucinations that violate graph topology.&lt;/p>
&lt;hr>
&lt;h2 id="method">Method&lt;/h2>
&lt;p>
&lt;figure >
&lt;div class="flex justify-center ">
&lt;div class="w-100" >&lt;img alt="Framework" srcset="
/blog/graphmaster/framework_hu1132089422936647416.webp 400w,
/blog/graphmaster/framework_hu5928412425996290846.webp 760w,
/blog/graphmaster/framework_hu2872404166475514373.webp 1200w"
src="https://enjundu.com/blog/graphmaster/framework_hu1132089422936647416.webp"
width="760"
height="257"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>GraphMaster decomposes graph synthesis into four specialized agents:&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Manager Agent&lt;/strong> — Selects between semantic and topological enhancement modes based on environmental analysis, and coordinates the entire synthesis workflow.&lt;/li>
&lt;li>&lt;strong>Perception Agent&lt;/strong> — Overcomes context-window limitations via semantic-aware community detection, mode-adaptive seed selection, and hierarchical PPR-based diffusion sampling to extract representative subgraphs.&lt;/li>
&lt;li>&lt;strong>Enhancement Agent&lt;/strong> — Generates new nodes and edges conditioned on extracted knowledge, with dual-mode generation for semantic coherence and structural fidelity.&lt;/li>
&lt;li>&lt;strong>Evaluation Agent&lt;/strong> — Assesses quality through multi-dimensional scoring (semantic + structural), with adaptive threshold and temporal convergence detection for iterative refinement.&lt;/li>
&lt;/ul>
&lt;hr>
&lt;h2 id="experimental-results">Experimental Results&lt;/h2>
&lt;p>Evaluated on &lt;strong>6 data-limited benchmarks&lt;/strong> with &lt;strong>4 GNN architectures&lt;/strong> (GCN, JKNET, GraphSage, GAT) on 8x A100 GPUs using QwQ-32B as the base LLM.&lt;/p>
&lt;p>
&lt;figure >
&lt;div class="flex justify-center ">
&lt;div class="w-100" >&lt;img alt="Main Results" srcset="
/blog/graphmaster/main_results_hu15599672716564817252.webp 400w,
/blog/graphmaster/main_results_hu17387086646069390212.webp 760w,
/blog/graphmaster/main_results_hu8138420897292343079.webp 1200w"
src="https://enjundu.com/blog/graphmaster/main_results_hu15599672716564817252.webp"
width="760"
height="269"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>GraphMaster consistently outperforms all baselines across all datasets. The bottom row (blue) shows GraphMaster achieving the highest accuracy and F1 scores on every benchmark.&lt;/p>
&lt;h3 id="graph-feature-preservation">Graph Feature Preservation&lt;/h3>
&lt;p>
&lt;figure >
&lt;div class="flex justify-center ">
&lt;div class="w-100" >&lt;img alt="Feature Analysis" srcset="
/blog/graphmaster/feature_analysis_hu9285242773121029112.webp 400w,
/blog/graphmaster/feature_analysis_hu7005581667582390789.webp 760w,
/blog/graphmaster/feature_analysis_hu10889657734607540595.webp 1200w"
src="https://enjundu.com/blog/graphmaster/feature_analysis_hu9285242773121029112.webp"
width="760"
height="269"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>The synthesized graphs maintain high fidelity: &lt;strong>KS statistic 0.357&lt;/strong> (p=0.059) for degree distribution, &lt;strong>0.835&lt;/strong> clustering coefficient similarity, and &lt;strong>0.988&lt;/strong> label homogeneity — indicating near-perfect structural preservation.&lt;/p>
&lt;h3 id="ablation-study">Ablation Study&lt;/h3>
&lt;p>
&lt;figure >
&lt;div class="flex justify-center ">
&lt;div class="w-100" >&lt;img alt="Ablation" srcset="
/blog/graphmaster/ablation_hu2146630035767259003.webp 400w,
/blog/graphmaster/ablation_hu7840005300923044651.webp 760w,
/blog/graphmaster/ablation_hu17319609184545855075.webp 1200w"
src="https://enjundu.com/blog/graphmaster/ablation_hu2146630035767259003.webp"
width="760"
height="257"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>Removing the Evaluation Agent causes the largest performance drop, confirming the critical role of iterative quality control. Each agent contributes uniquely to the final synthesis quality.&lt;/p></description></item><item><title>MoKGR: Mixture of Length and Pruning Experts for Knowledge Graphs Reasoning</title><link>https://enjundu.com/blog/mokgr/</link><pubDate>Sun, 19 Jan 2025 00:00:00 +0000</pubDate><guid>https://enjundu.com/blog/mokgr/</guid><description>&lt;h2 id="abstract">Abstract&lt;/h2>
&lt;p>Knowledge Graph (KG) reasoning critically depends on constructing informative reasoning paths. Existing GNNs adopt rigid, query-agnostic strategies. We propose &lt;strong>MoKGR&lt;/strong>, a mixture-of-experts framework with two innovations: &lt;strong>(1)&lt;/strong> a &lt;strong>mixture of length experts&lt;/strong> that adaptively weights path lengths based on query complexity, and &lt;strong>(2)&lt;/strong> a &lt;strong>mixture of pruning experts&lt;/strong> that evaluates paths from complementary perspectives. MoKGR achieves &lt;strong>state-of-the-art&lt;/strong> in both transductive and inductive settings.&lt;/p>
&lt;hr>
&lt;h2 id="motivation">Motivation&lt;/h2>
&lt;p>
&lt;figure >
&lt;div class="flex justify-center ">
&lt;div class="w-100" >&lt;img alt="Motivation" srcset="
/blog/mokgr/motivation_hu18331973836230186814.webp 400w,
/blog/mokgr/motivation_hu8395380476757186774.webp 760w,
/blog/mokgr/motivation_hu1535173140742592088.webp 1200w"
src="https://enjundu.com/blog/mokgr/motivation_hu18331973836230186814.webp"
width="760"
height="263"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>Consider two queries on the same graph: &lt;em>(JACK, followed, ?)&lt;/em> can be resolved within 3 hops, while &lt;em>(JACK, watched, ?)&lt;/em> requires deeper exploration. Existing methods use &lt;strong>fixed reasoning depth&lt;/strong> for all queries and &lt;strong>uniform pruning criteria&lt;/strong> — both are suboptimal. MoKGR personalizes both the depth and the pruning strategy per query.&lt;/p>
&lt;hr>
&lt;h2 id="method">Method&lt;/h2>
&lt;p>MoKGR introduces two complementary Mixture-of-Experts modules:&lt;/p>
&lt;ul>
&lt;li>&lt;strong>Mixture of Length Experts&lt;/strong> — Multiple experts specialized for different reasoning depths. A learned gating network computes query-specific weights over path lengths: simple queries route to short paths, complex queries activate deeper experts. The final output is a soft weighted combination.&lt;/li>
&lt;li>&lt;strong>Mixture of Pruning Experts&lt;/strong> — At each GNN layer, multiple pruning experts evaluate candidate paths from complementary perspectives (structural, semantic, diversity). A learned aggregation selects the top-k most informative paths per query.&lt;/li>
&lt;li>&lt;strong>End-to-End Training&lt;/strong> — Both modules are trained jointly with the answer prediction objective, ensuring optimal synergy between depth selection and path pruning.&lt;/li>
&lt;/ul>
&lt;hr>
&lt;h2 id="experimental-results">Experimental Results&lt;/h2>
&lt;h3 id="transductive-setting-6-benchmarks">Transductive Setting (6 Benchmarks)&lt;/h3>
&lt;p>
&lt;figure >
&lt;div class="flex justify-center ">
&lt;div class="w-100" >&lt;img alt="Main Results" srcset="
/blog/mokgr/main_results_hu4027277364006651311.webp 400w,
/blog/mokgr/main_results_hu18083655024233074811.webp 760w,
/blog/mokgr/main_results_hu9320919145034146317.webp 1200w"
src="https://enjundu.com/blog/mokgr/main_results_hu4027277364006651311.webp"
width="760"
height="377"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>MoKGR achieves the best results across all 6 benchmarks (Family, UMLS, WN18RR, FB15k237, NELL-995, YAGO3-10), outperforming both non-GNN baselines and state-of-the-art GNN methods including NBFNet, RED-GNN, A*Net, and AdaProp.&lt;/p>
&lt;h3 id="efficiency--convergence">Efficiency &amp;amp; Convergence&lt;/h3>
&lt;p>
&lt;figure >
&lt;div class="flex justify-center ">
&lt;div class="w-100" >&lt;img alt="Efficiency" srcset="
/blog/mokgr/efficiency_hu456109630554495659.webp 400w,
/blog/mokgr/efficiency_hu15649622221185549166.webp 760w,
/blog/mokgr/efficiency_hu12605689778133533305.webp 1200w"
src="https://enjundu.com/blog/mokgr/efficiency_hu456109630554495659.webp"
width="760"
height="343"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>MoKGR converges significantly faster than competing methods while maintaining lower inference time, demonstrating that personalized path exploration is both more accurate and more efficient.&lt;/p>
&lt;h3 id="expert-selection-analysis">Expert Selection Analysis&lt;/h3>
&lt;p>
&lt;figure >
&lt;div class="flex justify-center ">
&lt;div class="w-100" >&lt;img alt="Expert Analysis" srcset="
/blog/mokgr/expert_analysis_hu3306128304441321640.webp 400w,
/blog/mokgr/expert_analysis_hu15479781568752159853.webp 760w,
/blog/mokgr/expert_analysis_hu2064557752867909043.webp 1200w"
src="https://enjundu.com/blog/mokgr/expert_analysis_hu3306128304441321640.webp"
width="631"
height="760"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>The learned gating weights show interpretable patterns: the model prefers medium-length paths overall, but adapts dynamically — shorter paths for simple relational queries, deeper paths for complex multi-hop reasoning.&lt;/p>
&lt;h3 id="ablation-study">Ablation Study&lt;/h3>
&lt;p>
&lt;figure >
&lt;div class="flex justify-center ">
&lt;div class="w-100" >&lt;img alt="Ablation" srcset="
/blog/mokgr/ablation_hu12758446644151472906.webp 400w,
/blog/mokgr/ablation_hu6592786270217265840.webp 760w,
/blog/mokgr/ablation_hu10597442056394429424.webp 1200w"
src="https://enjundu.com/blog/mokgr/ablation_hu12758446644151472906.webp"
width="760"
height="640"
loading="lazy" data-zoomable />&lt;/div>
&lt;/div>&lt;/figure>
&lt;/p>
&lt;p>Removing the length experts causes the largest drop, confirming that adaptive depth is the most critical innovation. Removing pruning experts or using a single expert also degrades performance.&lt;/p></description></item></channel></rss>