Back to top papers

Unifying Ontology Construction and Semantic Alignment for Deterministic Enterprise Reasoning at Scale

Hongyin Zhu

Scoren/a
LLMn/a
Embedding0.407
Recencyn/a

Feedback

Why It Matters

No evaluation available.

Contributions

  • None available.

Insights

  • None available.

Limitations

  • None available.

Tags

  • cs.AI
  • cs.CL

Abstract

arXiv:2604.09608v1 Announce Type: new Abstract: While enterprises amass vast quantities of data, much of it remains chaotic and effectively dormant, preventing decision-making based on comprehensive information. Existing neuro-symbolic approaches rely on disjoint pipelines and struggle with error propagation. We introduce the large ontology model (LOM), a unified framework that seamlessly integrates ontology construction, semantic alignment, and logical reasoning into a single end-to-end architecture. LOM employs a construct-align-reason (CAR) pipeline, leveraging its unified architecture across all three stages: it first autonomously constructs a domain-specific ontological universe from raw data, then aligns neural generation with this structural reality using a graph-aware encoder and reinforcement learning, and finally executes deterministic reasoning over the constructed topology, node attributes and relation types. We evaluate LOM on a comprehensive benchmark constructed from diverse real-world enterprise datasets. Experimental results demonstrate that LOM-4B achieves 88.8% accuracy in ontology completion and 94% in complex graph reasoning tasks, significantly outperforming state-of-the-art LLMs. These findings validate that autonomous logical construction is essential for achieving deterministic, enterprise-grade intelligence.