Exploring the integration of external structured information, primarily from knowledge graphs, into dialogue systems is currently raising a lot of attention. Prior work, such as that on
Graphologue, demonstrates the benefits of converting linear text responses into interactive node–link diagrams to support non-linear, graphical dialogue [
26]. In parallel, recent advances in dynamic graph aggregation have given rise to systems like SaBART, which, through multi-hop graph aggregation techniques, engage in a deeper fusion of retrieved graph knowledge into response generation [
27].
3.1. Theoretical Framework of D-LLM
3.1.1. Conceptual Foundation
Dialogical Large Language Models (D-LLMs) initiate the perspective of a paradigmatic shift in AI architecture by coupling GRAPHYP’s structured graph-based reasoning with the natural language capabilities of Large Language Models. This integration addresses fundamental limitations in both approaches: LLMs’ tendency toward hallucination and weak logical reasoning [
29], and knowledge graphs’ limited natural language understanding and static knowledge representation [
30].
The theoretical foundation of D-LLM rests on three core principles:
Dialogical Intelligence: Moving beyond simple query–response interactions to sustained, contextual dialogues where the system maintains coherent reasoning across multiple conversational turns. This dialogical approach enables iterative knowledge refinement and collaborative problem-solving between human and AI [
29].
Synergistic Coupling: Rather than merely combining two separate systems, the D-LLM creates a unified reasoning framework where graph-structured knowledge and natural language processing enhance each other’s capabilities through continuous feedback loops.
Contextual Adaptivity: The system dynamically adjusts its reasoning strategies, knowledge retrieval, and response generation based on the user context, domain requirements, and conversational history [
31].
3.1.2. An Innovative Concept of Language Games as Foundational Theory
Central to the D-LLM’s theoretical framework is Wittgenstein’s concept of language games [
32]—the idea that language derives meaning from its use within specific social activities or “games” governed by contextual rules [
33]. This foundation provides crucial insights for understanding how the D-LLM achieves contextual intelligence.
Wittgenstein argued that words and sentences gain meaning only within particular language games—specific forms of language use embedded in social practices and activities, each with its own rules and purposes (e.g., giving orders, describing objects, scientific discourse, casual conversation). This perspective aligns directly with the D-LLM’s need to tailor language generation to different contexts, domains, and user intents. Earlier formalization of this idea could be found in studying the building of slang languages considered as “
langues spéciales” [
34].
In the D-LLM framework, GRAPHYP functions as a contextual mediator that identifies and models different “language games” by capturing the following:
Contextual parameters: Domain-specific vocabulary, discourse patterns, and communication norms;
User preferences: Individual communication styles, expertise levels, and interaction goals;
Social backgrounds: Professional contexts, cultural considerations, and community-specific language practices.
The D-LLM develops the perspective of operationalizing those insights through computational mechanisms:
Game Recognition Perspective: The system identifies, from the value of the three parameters of preferences recorded in GRAPHYP (Intensity, Variety, Attention), which “language game” a user is engaged in (technical consultation, educational dialogue, creative collaboration) and adjusts its linguistic behavior accordingly.
Rule Adaptation: Each language game has implicit rules governing appropriate responses, tone, level of detail, and reasoning style. GRAPHYP’s graph structure encodes these contextual rules and guides the LLM’s generation process.
Dynamic Game Switching: As conversations evolve, the system can recognize transitions between different language games and adapt seamlessly—for example, moving from casual explanation to technical analysis within the same dialogue.
3.1.3. Core Coupling Principles
The following four principles guide the D-LLM framework design. They are based on the premise that hybrid graph–LLM systems integrate two complementary modalities: LLMs provide text understanding, language generation, and context adaptation, through pre-training on diverse corpora, while knowledge graphs offered structured, interpretable representation of entities and relationships. In such systems, “cognitive communities” refer to clusters or networks within the graph structure that capture semantic, social, and relational connections among users, items, and contextual factors.
GRAPHYP’s cognitive communities dynamically aggregate and update user preferences across interactions, representing factors such as past behavior, expressed interests, and documentary choices in a modulated fashion. Integration with LLM reasoning capabilities enables conversational AI to generate responses that are contextually rich and anchored in an explicit, continuously updated user profile model [
11].
The four coupling principles are as follows:
D-LLM implements sustained dialogical interaction through iterative reasoning cycles where the LLM queries GRAPHYP for specific nodes, paths, or subgraphs, interprets the results, and refines subsequent queries based on previous answers. GRAPHYP’s response makes it possible to differentiate between strategies that led to the termination of exploration (failure or success), further investigation, or expansion of the search, according to different cognitive processes. This enables complex, multi-step reasoning tasks such as tracing relationships across several hops or synthesizing information from disparate parts of the knowledge graph. Advanced frameworks (e.g., Tree-of-Traversals [
35], GraphOTTER [
36]) empower the LLM to select discrete graph actions at each reasoning step.
In multi-turn conversations, the system maintains rich context about previous queries, answers, and reasoning paths. This contextual awareness enables follow-up questions, clarifications, and deeper exploration of the knowledge graph while preserving conversational coherence.
Unlike black-box AI systems, D-LLM constructs explicit, interpretable reasoning traces. The LLM selects discrete graph actions (such as VisitNode, GetSharedNeighbours, or AnswerQuestion) at each reasoning step, creating clear audit trails essential for transparency and explainability.
By anchoring each reasoning step in the actual graph structure, D-LLM reduces hallucinations and ensures factual accuracy. This grounding is particularly crucial for multi-hop queries and knowledge-intensive tasks where precision is paramount.
3.1.4. Dialogical vs. Traditional Approaches: D-LLM Perspective
Traditional AI systems typically operate through isolated query–response cycles with limited context retention. D-LLM’s dialogical approach enables the following:
Conversational Memory: The system builds comprehensive models of ongoing dialogues, tracking not just facts exchanged but reasoning patterns, user preferences, and evolving understanding.
Collaborative Discovery: Rather than simply retrieving pre-existing knowledge, D-LLM engages in collaborative knowledge construction, helping users explore ideas, test hypotheses, and develop insights through sustained interaction.
Adaptive Expertise: The system adjusts its level of explanation, terminology, and reasoning depth based on demonstrated user expertise and feedback, creating truly personalized learning experiences.
Figure 1 sums up the characteristics that we propose for the D-LLM:
A prospective example of D-LLM, in
Appendix A, illustrates potential dialogical interactions between GRAPHYP and an LLM, while the GRAPHYP perspective on how to resolve scientific disputes with LLMs is illustrated in
Appendix B.
3.1.5. Methodological Implications: The Unique Advantages of Human Preference Expression
The D-LLM framework offers several important advantages through its integration of GRAPHYP’s cognitive communities:
Nuanced Representation of Complex Preferences. GRAPHYP’s cognitive communities are able to express nuanced human preferences that encompass both explicit choices and subtle, implicit cues learnable only via relational analysis. Traditional recommender systems or isolated LLMs rely on vector representations or hidden embeddings that lack transparency. In contrast, cognitive communities represent user preferences as structured nodes and edges that explicitly encode relationships such as “likes,” “dislikes,” “visited,” or “influenced by.” This approach enables complex multi-hop relations among diverse data points while allowing community detection that reveals shared interests and collective biases among groups of users.
Enhanced Multi-hop Reasoning and Context Aggregation. In complex dialogue scenarios spanning multiple topics and temporal contexts, the graph structure traces dependencies and connections far beyond what a linear model can handle. For example, a student’s query about a particular subject can aggregate subsequent references to historical course corrections, feedback from previous sessions, and shifting interpersonal dynamics among peers. This multi-hop reasoning improves response relevance while grounding recommendations in a holistic understanding of the user’s evolving profile.
Transparency and Explainability. Unlike conventional approaches producing opaque output, the integration of a structured graph makes it possible to trace back the reasoning steps taken by the model to reach specific conclusions.
Collaborative and Community-driven Personalization. Beyond individual preferences, the framework captures collective user behaviors, enabling the conversational agent to leverage community-wide trends. Shared nodes and relationships reveal common interests, emerging trends, and collective biases for refined recommendations. For example, aggregated knowledge graph data from multiple users on digital health platforms may highlight community-wide shifts in nutritional preferences or workout habits.
Ethical Considerations and Privacy Preservation. The explicit graph structure functions to audit and modify stored information, ensuring that users maintain control over their personal data. Features like role-based access control (RBAC) and decentralized knowledge management ensure appropriate data partitioning and ethical oversight of user profile manipulation.
These advantages enable the D-LLM to achieve three key capabilities:
Beyond Information Retrieval: The framework transcends traditional information retrieval by enabling genuine knowledge co-construction through dialogue. Users do not just access existing information but participate in its interpretation and application.
Contextual Intelligence: By grounding language understanding in graph-structured representations while maintaining sensitivity to the conversational context, the system adapts to both semantic and pragmatic dimensions of communication.
Human–AI Collaboration: The approach positions AI not as a replacement for human reasoning but as a sophisticated cognitive tool that amplifies human capabilities while preserving human agency and choice.
This foundation provides the D-LLM with a novel perspective on human–AI interaction, combining the precision of structured representation with the flexibility and naturalness of conversational AI.
3.5. Applications and Use Cases
The D-LLM framework translates theoretical capabilities into practical applications across diverse domains by integrating GRAPHYP’s graph-based reasoning with LLM natural language processing.
3.5.1. Scientific Research and Knowledge Discovery
Dispute Resolution and Conflict Analysis: GRAPHYP’s dispute learning visualizes conflicting scientific claims as graph structures, mapping opposing claims, supporting evidence, and connecting pathways [
17]. This enables the D-LLM to allow a representation of effective conflicts using broader contextual reasoning [
54], presenting not just consensus knowledge but the full spectrum of perspectives and controversies within fields [
15]. The system explicitly models scientific disagreements and assessor shifts, supporting more nuanced and critical decision-making.
Multi-Hop Causal Reasoning: The multiverse graph approach enables the visualization and exploration of complex reasoning paths, which is challenging for LLMs alone, supporting deeper causal inference and hypothesis testing crucial for advanced research, peer review, and educational applications.
3.5.2. Personalized Learning and Education
Modern personalized learning systems increasingly leverage hybrid architectures that combine graph-based knowledge representation with large language model capabilities. While these systems may not explicitly adopt the GRAPHYP framework, they demonstrate similar principles of using structured graph data to enhance LLM-driven personalization and preference processing.
Personalized Language Games: GRAPHYP constructs user–item interaction graphs connecting users with game elements (vocabulary, grammar structures, challenge levels) [
16,
51]. The system enables dynamic difficulty adjustment through action sequence analysis, personalized hint systems identifying deviation points from successful pathways, and branching narrative generation through adversarial clique detection in player choice patterns. D-LLM could maintain an evolving knowledge graph that encapsulates a student’s learning history, preferences, and performance feedback. Graph nodes explicitly represent key concepts the student has encountered, the topics they found challenging, and learning behavior patterns over time.
Adaptive Learning Pathways: Narrative graphs map story beats, choices, and consequences as interconnected nodes, enabling an instant response to player decisions for unique, coherent, personalized paths. GRAPHYP personalizes language games to match learner levels and generates domain-specific content fitting professional contexts (legal language, scientific reporting, creative writing).
Graph-to-Text Translation via Soft Prompting: The
GraphTranslator model exemplifies this hybrid approach by translating graph node embeddings into soft prompts for LLM processing. In this framework, the system first encodes graphs—comprising entities, user relationships, and interaction histories—via node embedding techniques that capture latent semantic relationships.
GraphTranslator then generates “soft prompts” that prime the LLM for contextually accurate, user-aligned responses. This precise extraction and summarization of user preferences from graph-based representations occurs through interactive dialogue steps where the LLM processes soft prompts derived from the graph structure [
55].
Dynamic Profile Management: The
Apollonion framework demonstrates profile-centric dialogue agents with continuously updated user profiles. Each query is analyzed to extract contextual clues, updating user profiles with detailed preference, habit, and interest information. Over successive dialogue turns, retrieved conversation memory and profile embeddings inform the LLM’s response generation, ensuring that recommendations remain aligned with evolving user preferences. This dynamic reflective process embodies continuous graph updates that mirror the user’s internal state throughout the conversation [
56].
Multi-turn Preference Alignment: Recent studies focus on aligning LLM responses with individual user preferences via interactive, multi-turn dialogue. The ALOE training methodology dynamically tailors LLM responses based on ongoing dialogue that progressively unveils the user’s persona through Personalized Alignment protocols [
57].
Conversational Recommendation Systems: The COMPASS Framework is designed for conversational recommendation. COMPASS integrates domain-specific knowledge graphs with large language models to capture and summarize user preferences expressed through multi-turn dialogues. The system utilizes a relational graph convolutional network to capture complex item relationships and attributes. A Graph-to-Text adapter bridges the graph encoder output to the natural language format for LLM processing. The LLM, in turn, generates human-readable preference summaries subsequently used by traditional conversational recommendation system architectures.
Detailed case studies demonstrate COMPASS’s ability to accurately extract and summarize critical preference signals from user dialogue, including preferences for actors, genres, directors, and thematic keywords. Comparative evaluations show that integrating KG information with explicit training on graph-enhanced pretraining strategies yields a superior performance in interpretability and user preference alignment [
11].
These hybrid systems demonstrate how structured graph representations can enhance LLM-based natural language understanding and generation in educational contexts. The integration of graph-encoded user data with conversational AI creates more nuanced, adaptive learning experiences that respond to individual learning patterns and preferences while maintaining pedagogical effectiveness across diverse educational domains.
3.5.3. Content Verification and Fact-Checking
Enhanced Verification Framework: GRAPHYP’s causal-first knowledge graphs provide LLMs with explicit, verified relationships during text generation, enabling real-time cross-referencing against factual nodes. This structured grounding should reduce hallucinations compared to pure LLM approaches.
Explainable Fact-Checking: The system embodies Explainable AI principles through the following:
Reasoning Path Traversal: Step-by-step visualization from input to output;
Entity Linking and Source Tracing: Connecting text mentions to uniquely identified entities for semantic annotation and provenance tracking;
Dispute Modeling: Surfacing alternative reasoning paths and highlighting uncertainty in conflicting or ambiguous cases.
Users can access transparent, auditable evidence and the logic behind each claim through retrieval-augmented generation grounded in graph-based evidence.
3.5.4. Interactive Systems and Dialogue Applications
Context-Aware Dialogue: GRAPHYP identifies the user’s language game (casual chat, technical support, educational tutoring) and steers the LLM to adopt corresponding patterns and tone. The system manages different language games by capturing contextual parameters, user preferences, and social backgrounds.
Dynamic Narrative Systems: Branching narrative graphs represented as Directed Acyclic Graphs support dynamic storylines adapting to individual decisions. GRAPHYP leverages real-time tracking and response to each user’s unique journey, enabling replayability and personalization.
Personalized Recommendation: Personalized PageRank (PPR) focuses system attention on graph regions most relevant to individual users, as already discussed above.
3.5.5. Advanced Reasoning and Decision Support
Real-Time Knowledge Integration: Knowledge Graph Tuning (KGT) allows LLMs to update knowledge bases using structured GRAPHYP inputs without costly retraining. Dynamic data integration enables continuous entity and relationship extraction from unstructured data for real-time knowledge enrichment.
Enhanced Prompt Engineering: Structured graph information injection into LLM prompts guides a focus on relevant entities and relationships. Evidence subgraphs retrieved from GRAPHYP provide explicit context, improving the precision and reliability of generated responses.
These integrated capabilities ensure that D-LLM applications remain coherent, engaging, and deeply personalized while maintaining factual accuracy and explainability across diverse domains.
3.6. Human Choice Freedom and Preference Expression
The coupling of GRAPHYP and LLMs in D-LLM fundamentally transforms how users interact with AI systems by establishing a human-choice-first framework that expands the dimensions of preference expression and knowledge discovery. Rather than constraining human agency, this integration enhances human free choice by improving the accuracy, reliability, and interpretability of AI outputs, empowering humans to make more informed and autonomous decisions while leveraging AI as a complementary cognitive tool rather than a replacement.
Integrating graph structures with LLMs enhances the AI’s ability to perform complex, multi-step reasoning [
54]. Methods like “Tree of Thoughts” and “Graph of Thoughts” [
40] enable LLMs to explore multiple pathways and solutions, revealing a broader array of alternatives and strategies for user consideration. This approach ensures that users are presented with more diverse and creative options, not just the most obvious or common ones, thereby expanding their decision-making landscape beyond the limitations of traditional AI systems.
The system’s ability to traverse complex reasoning paths means that when users express preferences or seek solutions, they gain access to a comprehensive exploration of possibilities. This enhanced decision-making capability operates through the synergistic combination of GRAPHYP’s structured knowledge representation and the LLM’s natural language understanding, creating a collaborative environment where human creativity and AI capability enhance each other.
Coupling GRAPHYP with LLMs revolutionizes preference modeling through several key mechanisms. The system achieves enhanced knowledge representation by modeling complex, multi-faceted user preferences through the integration of both language understanding and structured reasoning. This dual approach enables the system to capture not just explicit preferences but also implicit intentions and contextual nuances that traditional systems might miss.
Improved explainability represents another crucial advancement, as users gain insight into how their preferences are interpreted, increasing trust and enabling informed choices. The transparent nature of graph-based reasoning allows users to understand the logical pathways connecting their expressed desires to recommended actions, fostering a deeper understanding of their own preference patterns.
Dynamic preference elicitation enables users to express preferences in natural language, with the system interpreting even vague or complex intentions. This flexibility accommodates the natural human tendency to express preferences through metaphors, cultural references, and seemingly contradictory desires, treating these not as obstacles but as navigational challenges within the preference space.
The D-LLM method achieves comprehensive personalization through three fundamental principles that preserve and enhance human agency. Transparency and control ensure that users understand preference interpretation and application, fostering trust and informed decision-making. Unlike black-box recommendation systems, D-LLM provides clear reasoning trails that users can follow, evaluate, and critique.
Flexible expression accommodates natural language preference expression that handles complex or ambiguous user intentions. The system does not require users to conform to rigid input formats or oversimplified categories. Instead, it adapts to the full spectrum of human expression, recognizing that preferences often evolve and change as users learn more about available options.
Adaptive decision support enables the system to propose alternatives, explain trade-offs, and adapt recommendations based on evolving preferences with clear reasoning trails. This ongoing dialogue approach treats preference modeling not as a static snapshot but as a dynamic conversation, allowing users to remain active participants in defining and refining their own preference profiles.
GRAPHYP can rely on several strategies to handle scaling with large user bases and frequent, real-time graph updates:
Data Partitioning and Sharding: The system achieves horizontal scaling through sharding, dividing large graphs into smaller, manageable subgraphs distributed across different machines or clusters. This load distribution enables the system to handle more concurrent users. Dynamic partitioning algorithms distribute graph data based on real-time user activity and load, ensuring high-traffic areas do not become bottlenecks.
Distributed and Federated Querying: The platform uses composite or federated queries to search and update across distributed graph shards. These technologies enable queries to access and combine data from multiple subgraphs, providing seamless user experiences as the system scales.
Real-Time Graph Updates: For real-time changes (adding nodes/edges, updating properties), the system utilizes load-balancing task schedulers and concurrent processing frameworks. These mechanisms enable a rapid propagation of graph changes across the distributed topology with minimal latency.
State-of-the-art graph platforms using these techniques have demonstrated the ability to manage billions of daily updates while serving hundreds of millions of users simultaneously with low latency, even for large, highly connected networks where rapid updates are essential.
The transformation from traditional LLM approaches to the D-LLM’s integrated framework represents a fundamental shift in how AI systems handle human preferences and choice. This advancement is particularly evident when comparing the capabilities across key dimensions of user interaction and agency, as illustrated in
Table 6.
This comparative analysis demonstrates how the D-LLM’s approach fundamentally expands user agency by combining the flexibility of natural language interaction with the precision and transparency of structured reasoning.
The integration ensures that choices are better understood, accurately modeled, and dynamically updated according to user needs and context. This human-choice-first framework operates on the principle that AI should amplify rather than replace human decision-making capabilities, creating a collaborative cognitive environment where users maintain agency while benefiting from enhanced information access, expanded option awareness, and transparent reasoning support.
Through this approach, the D-LLM establishes a new paradigm for human–AI interaction—one that preserves human autonomy while providing powerful cognitive augmentation, ensuring that the ultimate goal remains empowering humans to make better choices for themselves rather than having choices made for them by algorithmic systems.