Relationship Mapping Techniques
Relationship Mapping Techniques in AI Discoverability Architecture represent systematic methodologies for identifying, visualizing, and leveraging connections between entities, concepts, and data structures to enhance the findability and accessibility of AI systems and their outputs 1. These techniques serve as foundational mechanisms that enable AI systems to understand semantic relationships, contextual dependencies, and hierarchical structures within complex information ecosystems 2. In the context of AI discoverability, relationship mapping is critical because it transforms unstructured or semi-structured data into navigable knowledge graphs that facilitate intelligent search, recommendation, and retrieval operations 3. As AI systems become increasingly integrated into enterprise architectures and consumer applications, the ability to map and traverse relationships efficiently determines the effectiveness of discovery mechanisms and the overall user experience.
Overview
The emergence of Relationship Mapping Techniques stems from the fundamental limitation of traditional keyword-based search and retrieval systems, which fail to capture the semantic richness and contextual dependencies inherent in complex information landscapes 12. As organizations accumulated vast repositories of unstructured and semi-structured data throughout the early 2000s, the inadequacy of simple text matching became increasingly apparent, driving the development of more sophisticated approaches that could understand and leverage the connections between information elements 3.
The fundamental challenge these techniques address is transforming isolated data points into interconnected knowledge structures that support intelligent discovery and reasoning 4. Traditional search systems treat documents and data as independent units, missing the valuable insights that emerge from understanding how entities relate to one another. For instance, knowing that a particular research paper cites another, that an author works at a specific institution, and that the institution collaborates with certain companies creates a rich web of connections that enables more sophisticated discovery than simple keyword matching 5.
The practice has evolved significantly from early rule-based systems and manually curated ontologies to modern neural approaches that can automatically extract and learn relationship representations from large-scale datasets 67. Contemporary implementations leverage transformer-based architectures, graph neural networks, and embedding techniques that represent relationships as vectors in high-dimensional spaces, enabling mathematical operations on semantic connections 8. This evolution has been driven by advances in machine learning, the availability of large training datasets, and the computational resources necessary to process billions of entities and relationships at scale 9.
Key Concepts
Entity Recognition and Linking
Entity recognition and linking involves identifying discrete objects or concepts within data sources and connecting them to canonical representations in a knowledge base 1. This process addresses the challenge that the same entity may be referenced in multiple ways across different documents or data sources, requiring disambiguation to ensure all mentions refer to the correct real-world object.
For example, in a pharmaceutical research context, the entity "aspirin" might appear as "acetylsalicylic acid," "ASA," or various brand names across different research papers and clinical databases. An entity recognition and linking system would identify all these mentions, recognize they refer to the same compound, and link them to a canonical identifier in a chemical database like PubChem. This enables researchers searching for information about aspirin to discover all relevant documents regardless of which terminology was used, while also distinguishing this compound from other entities that might share similar names.
Knowledge Graph Construction
Knowledge graph construction transforms extracted entities and relationships into structured graph representations where nodes represent entities and edges represent the connections between them 23. These graphs implement formal schemas that define entity types, relationship types, and the properties associated with each, enabling efficient querying and traversal operations.
Consider a media company building a knowledge graph for content recommendation. The graph would contain nodes representing movies, actors, directors, genres, and viewers, with edges capturing relationships like "acted_in," "directed," "belongs_to_genre," and "watched." A specific instance might show that the node for actor "Meryl Streep" connects via an "acted_in" edge to the node for "The Devil Wears Prada," which connects via a "directed_by" edge to "David Frankel" and via a "belongs_to_genre" edge to "Comedy-Drama." This structured representation enables the system to answer complex queries like "find comedy-dramas directed by people who have worked with Meryl Streep" by traversing the appropriate relationship paths.
Semantic Reasoning and Inference
Semantic reasoning and inference capabilities allow systems to derive implicit relationships from explicit ones, expanding the relationship network beyond directly observed connections 45. This component implements ontological reasoning, rule-based inference, or neural reasoning mechanisms that apply logical rules or learned patterns to infer new knowledge.
In a healthcare knowledge graph, explicit relationships might indicate that "Drug A treats Condition X" and "Condition X is a type of Inflammatory Disease." Through semantic reasoning, the system can infer that "Drug A treats Inflammatory Disease" even if this relationship was never explicitly stated in the source data. More sophisticated reasoning might combine multiple relationship paths: if "Patient Population Y has high prevalence of Condition X" and "Drug A treats Condition X," the system might infer that "Drug A is potentially relevant for Patient Population Y," supporting clinical decision-making and drug discovery efforts.
Relationship Extraction and Classification
Relationship extraction and classification analyzes the context surrounding entity pairs to determine relationship types and strengths 67. Modern approaches employ neural relation extraction models that use attention mechanisms to identify relationship types such as "is-a," "part-of," "located-in," or domain-specific relationships, producing typed relationships with confidence scores.
For instance, in processing the sentence "Apple CEO Tim Cook announced the new iPhone at the company's Cupertino headquarters," a relationship extraction system would identify multiple entities (Apple, Tim Cook, iPhone, Cupertino) and extract several relationships: "Tim Cook" has an "employed_by" relationship with "Apple" (with high confidence based on the "CEO" context), "iPhone" has a "manufactured_by" relationship with "Apple," and "Apple" has a "headquartered_in" relationship with "Cupertino." Each extracted relationship would include a confidence score based on the strength of the contextual evidence, allowing downstream systems to prioritize more reliable connections.
Graph Embeddings
Graph embeddings represent entities and relationships as vectors in continuous high-dimensional spaces, enabling mathematical operations on semantic connections 89. Techniques like TransE model relationships as translations in embedding space, while more sophisticated approaches use rotations in complex or quaternion spaces to capture different relationship properties.
In a product recommendation system, graph embeddings might represent products, categories, and customer segments as vectors where semantically similar items are positioned close together in the embedding space. If the embedding for "wireless headphones" is positioned near "bluetooth speakers" and "portable audio," the system can identify these products as related even without explicit relationship labels. When a customer purchases wireless headphones, the system can compute vector similarities to recommend bluetooth speakers, leveraging the learned embedding structure. These embeddings also support link prediction—if the vector arithmetic suggests a relationship should exist between two entities but none is recorded, the system can flag this as a potential missing connection for validation.
Multi-Hop Reasoning
Multi-hop reasoning enables systems to answer complex queries by traversing multiple relationship paths through the knowledge graph 45. This capability is essential for questions that cannot be answered by examining a single relationship but require connecting information across several steps.
Consider a financial compliance system that needs to identify potential conflicts of interest. A single-hop query might find that "Executive A serves on the board of Company X." However, detecting a conflict of interest might require multi-hop reasoning: "Executive A serves on the board of Company X" → "Company X owns subsidiary Company Y" → "Company Y has a supply contract with Company Z" → "Company Z is being evaluated for acquisition by Executive A's primary employer." This four-hop path reveals a potential conflict that would be invisible to simpler query approaches, demonstrating how multi-hop reasoning uncovers non-obvious connections critical for compliance, fraud detection, and risk assessment.
Temporal Relationship Modeling
Temporal relationship modeling captures how relationships change over time, recognizing that many connections are not static but evolve, emerge, or dissolve 23. This involves timestamping relationships and implementing query mechanisms that can retrieve the state of the relationship graph at specific points in time or track relationship evolution.
In an organizational knowledge graph tracking company structure, temporal modeling would record that "Employee John Smith" had a "reports_to" relationship with "Manager Jane Doe" from January 2020 to June 2022, then this relationship ended and a new "reports_to" relationship with "Manager Robert Chen" began in July 2022. This temporal information enables queries like "who reported to Jane Doe during Q2 2021?" or "trace the reporting chain for this project that ran from 2020-2023," providing accurate historical context. Temporal modeling is particularly critical in domains like regulatory compliance, where understanding the state of relationships at specific historical moments is legally significant.
Applications in AI Discoverability Architecture
Intelligent Search Enhancement
Relationship mapping techniques dramatically enhance search capabilities by enabling semantic search that goes beyond keyword matching to understand conceptual connections 12. When a user searches for "machine learning applications in healthcare," a relationship-aware system leverages the knowledge graph to identify connections between ML techniques (neural networks, decision trees, clustering algorithms), medical conditions (diabetes, cancer, cardiovascular disease), and treatment outcomes (diagnosis accuracy, treatment efficacy, patient survival rates). The system can return relevant results even when these exact terms don't co-occur in documents, by traversing relationship paths that connect the query concepts to relevant content. For example, a research paper discussing "deep learning for diabetic retinopathy detection" would be surfaced because the system understands that deep learning is a type of machine learning, diabetic retinopathy is a healthcare condition, and detection relates to diagnosis—all connections captured in the relationship map.
Recommendation Systems
Relationship mapping enables sophisticated recommendation systems that traverse graph structures to identify relevant suggestions through multiple relationship paths 35. E-commerce platforms like Amazon use relationship graphs connecting products, customer preferences, purchase patterns, product categories, and contextual attributes. When a customer views a particular camera, the system doesn't simply recommend other cameras; instead, it traverses relationships to suggest complementary items: "customers who bought this camera" → "also purchased these lenses" → "and these camera bags" → "and these memory cards." The system can also identify less obvious recommendations by following longer paths: "this camera is used for wildlife photography" → "wildlife photographers also use telephoto lenses" → "users of telephoto lenses often need tripods with specific stability ratings." This multi-path traversal generates diverse, contextually relevant recommendations that simple collaborative filtering would miss.
Question Answering Systems
Knowledge graph-based question answering systems leverage relationship mapping to construct answers requiring multi-hop reasoning 46. Google's search engine employs relationship maps to answer complex queries like "Who directed the movie starring the actor who won the Best Actor Oscar in 2020?" The system parses this query to identify the reasoning chain required: first, identify which actor won Best Actor in 2020 (Joaquin Phoenix) → then find which movie he starred in that year (Joker) → then identify who directed that movie (Todd Phillips). Each step involves traversing specific relationship types in the knowledge graph: "won_award" relationships connecting actors to awards and years, "acted_in" relationships connecting actors to films, and "directed_by" relationships connecting films to directors. The relationship map enables the system to execute this reasoning chain and return "Todd Phillips" as the answer, demonstrating how relationship mapping transforms complex natural language questions into graph traversal operations.
Enterprise Knowledge Discovery
Organizations implement relationship mapping to structure internal knowledge as interconnected concepts, enabling knowledge discovery, expertise location, and decision support 78. A pharmaceutical company might build a knowledge graph connecting research publications, patents, clinical trials, researchers, compounds, diseases, and regulatory filings. When a scientist investigates potential treatments for a rare disease, the relationship map enables discovery of non-obvious connections: "Researcher A published on Compound X" → "Compound X showed efficacy in Trial Y for Disease Z" → "Disease Z shares a biological pathway with our target rare disease" → "Researcher A has expertise relevant to our investigation." This relationship-driven discovery surfaces relevant expertise and prior research that keyword search would miss, accelerating innovation by connecting distributed knowledge across the organization.
Best Practices
Implement Confidence Scoring and Provenance Tracking
Every extracted relationship should include a confidence score indicating the reliability of the connection, along with provenance information tracing the relationship to its source 16. This practice enables downstream systems to prioritize high-confidence relationships and allows human reviewers to validate uncertain connections. The rationale is that automated extraction methods inevitably produce errors, and making uncertainty explicit prevents low-quality relationships from degrading system performance.
For implementation, when a neural relationship extraction model identifies that "Company A acquired Company B" from a news article, the system should store not only the relationship but also a confidence score (e.g., 0.92 based on the model's output probability), the source document identifier, the specific text span that evidenced the relationship, and the extraction timestamp. This metadata enables queries like "show me all acquisition relationships with confidence above 0.85" or "find all relationships extracted from this particular source," supporting quality control workflows where human experts review low-confidence relationships before they're used in production applications.
Design Flexible Schemas with Evolution in Mind
Relationship schemas should accommodate extension and evolution without requiring complete restructuring as domains change and new relationship types emerge 27. The rationale is that knowledge domains are not static—new concepts, entity types, and relationship types continuously emerge, and rigid schemas become obsolete or require costly migrations.
Implement this by adopting property graph models that allow arbitrary properties on nodes and edges, rather than strictly enforced schemas that require predefined relationship types. For example, when building a knowledge graph for emerging technology domains, define core relationship types (is-a, part-of, related-to) but allow custom relationship types to be added dynamically. When a new relationship type like "competes-with-in-market-segment" becomes relevant, the system should accommodate it without schema migration. Maintain versioning for relationship type definitions, document semantic meanings in an ontology registry, and implement backward-compatible query patterns that gracefully handle schema evolution.
Establish Human-in-the-Loop Validation Workflows
Critical relationships should undergo human validation before being used in high-stakes applications, particularly in regulated domains like healthcare and finance 58. The rationale is that while automated extraction achieves high accuracy on average, individual errors can have serious consequences, and human expertise remains essential for validating complex or ambiguous relationships.
Implement validation workflows that route low-confidence relationships (below a domain-specific threshold) to subject matter experts for review. For instance, in a clinical knowledge graph, relationships extracted with confidence below 0.90 that connect drugs to side effects should be reviewed by pharmacologists before being used in clinical decision support systems. Create validation interfaces that show the relationship in context, display the source evidence, and allow experts to confirm, reject, or modify the relationship. Track validation decisions to create training data for improving extraction models, and implement feedback loops where validated relationships are used to retrain models, progressively reducing the validation burden.
Optimize for Query Patterns and Access Patterns
Graph database indexing and partitioning strategies should align with anticipated query patterns to ensure acceptable performance at scale 39. The rationale is that graph traversal performance degrades rapidly with poor indexing, and different query patterns require different optimization strategies.
For implementation, analyze the most common query patterns in your application—for example, if most queries start from user entities and traverse "purchased" relationships to products, then traverse "similar-to" relationships to find recommendations, create indexes optimized for this pattern. Partition large graphs by entity type or relationship type based on access patterns: if product-to-product relationships are frequently traversed together, co-locate these subgraphs. For a knowledge graph supporting research discovery, if queries typically start with author entities and traverse "authored" relationships to papers, then "cites" relationships to other papers, create composite indexes on (author, authored, paper) and (paper, cites, paper) paths. Monitor query performance continuously and adjust indexing strategies as usage patterns evolve.
Implementation Considerations
Graph Database Selection and Configuration
Choosing the appropriate graph database technology depends on scale requirements, query complexity, consistency needs, and integration constraints 23. Neo4j provides a mature property graph model with the Cypher query language, suitable for applications requiring ACID transactions and complex pattern matching. Amazon Neptune offers a managed service supporting both property graphs and RDF triple stores, appropriate for cloud-native applications requiring high availability. Apache Jena provides RDF and SPARQL support for applications requiring semantic web standards compliance.
For a mid-sized enterprise implementing a knowledge graph with millions of entities and tens of millions of relationships, Neo4j might be appropriate, configured with sufficient memory to cache frequently accessed graph regions (the "hot set") and SSD storage for the full graph. Index all properties used in query filters, configure relationship type indexes for common traversal patterns, and implement query result caching for frequently executed patterns. For larger scales exceeding billions of relationships, consider distributed graph databases like JanusGraph or implement custom sharding strategies that partition the graph across multiple database instances based on entity types or relationship communities.
Extraction Model Selection and Customization
The choice between rule-based, supervised learning, distant supervision, and neural extraction approaches depends on data characteristics, domain specificity, and available training resources 67. Rule-based extraction works well for highly structured domains with consistent patterns, while neural approaches excel with varied natural language but require substantial training data.
For a specialized domain like legal contract analysis, implement a hybrid approach: use rule-based extraction for standardized relationship types with consistent linguistic patterns (e.g., "Party A agrees to pay Party B" always indicates a payment obligation relationship), and employ fine-tuned transformer models for more variable relationships. Start with a pre-trained model like BERT or RoBERTa, then fine-tune on a domain-specific dataset of annotated legal documents. For domains with limited labeled data, use distant supervision by leveraging existing structured databases—for example, if you have a database of known company acquisitions, use it to automatically generate training examples by finding sentences mentioning both companies around the acquisition date, assuming these sentences express acquisition relationships.
Integration with Existing Enterprise Systems
Relationship mapping implementations must integrate with existing data sources, applications, and workflows 89. This requires ETL pipelines that extract data from diverse sources (databases, document repositories, APIs), transform it into graph structures, and load it into the graph database, while maintaining synchronization as source data changes.
For an organization with customer data in a CRM system, product data in an ERP system, and interaction data in web analytics platforms, implement a data integration pipeline using tools like Apache Airflow or AWS Glue. Schedule regular extraction jobs that pull updated data from each source, apply entity resolution to match customer entities across systems (using fuzzy matching on names and exact matching on email addresses), extract relationships (customer purchased product, customer viewed product, customer contacted support about product), and update the knowledge graph incrementally. Implement change data capture where possible to identify only modified records, reducing processing overhead. Expose the knowledge graph through APIs that existing applications can query, enabling gradual adoption without requiring wholesale application rewrites.
Privacy and Access Control Implementation
When relationship maps contain sensitive information, implement fine-grained access controls that restrict relationship visibility based on user roles and data sensitivity 5. The challenge is that relationships themselves may reveal confidential information even when individual entities are not sensitive—for example, knowing that a particular employee is connected to a specific project might reveal confidential strategic initiatives.
Implement relationship-level access controls by tagging relationships with sensitivity classifications and user roles authorized to view them. For a healthcare knowledge graph, relationships connecting patients to diagnoses should only be visible to authorized clinical staff, while relationships connecting diagnoses to treatment protocols might be visible to researchers. Use attribute-based access control (ABAC) policies that evaluate user attributes, relationship attributes, and contextual factors (time, location, purpose) to make access decisions. Implement query rewriting that automatically filters relationship traversals based on the requesting user's permissions, ensuring that graph queries only return relationships the user is authorized to see. Maintain audit logs of all relationship access for compliance and security monitoring.
Common Challenges and Solutions
Challenge: Entity Resolution Across Heterogeneous Sources
When integrating relationship maps from multiple sources with different schemas, granularities, and quality levels, determining when entities in different sources refer to the same real-world object remains a difficult problem 12. Different systems may use different identifiers, naming conventions, and levels of detail. For example, one system might reference "IBM Corporation" while another uses "International Business Machines," and a third uses stock ticker "IBM"—all referring to the same entity. Without accurate entity resolution, the relationship map fragments into disconnected subgraphs, missing valuable cross-source connections.
Solution:
Implement a multi-stage entity resolution pipeline combining deterministic matching, probabilistic matching, and machine learning approaches 3. First, apply deterministic rules for high-confidence matches: exact matches on unique identifiers (email addresses, government IDs, product SKUs) automatically link entities. Second, apply probabilistic matching using similarity metrics on entity attributes: if two company entities have Levenshtein distance below a threshold on name, matching addresses, and matching industry codes, they likely represent the same entity. Third, train a machine learning classifier on labeled examples of entity pairs (matched/not matched) using features like string similarity, attribute overlap, and contextual information. For the IBM example, the system would match "IBM Corporation," "International Business Machines," and "IBM" by recognizing that they share the same stock ticker, headquarters address, and industry classification, even though the names differ. Maintain a canonical entity registry that assigns unique identifiers to resolved entities, and create mapping tables that link source-specific identifiers to canonical identifiers. Implement human review workflows for ambiguous cases where automated matching confidence is below a threshold, and use validation decisions to improve matching models over time.
Challenge: Scalability with Billions of Relationships
As relationship maps grow to billions of entities and relationships, query performance degrades, storage costs increase, and maintenance operations become prohibitively expensive 79. A knowledge graph for a large e-commerce platform might contain billions of product entities, customer entities, and interaction relationships, making simple graph traversals that complete in milliseconds on small graphs take minutes or fail entirely on the full-scale graph.
Solution:
Implement a multi-tier architecture that combines hot/warm/cold storage strategies with intelligent caching and query optimization 8. Store frequently accessed relationships (the "hot set") in memory-optimized graph databases with SSD backing, while archiving historical or rarely accessed relationships to cheaper storage tiers. Implement graph partitioning strategies that co-locate frequently traversed subgraphs: for the e-commerce example, partition by product category so that queries exploring "customers who bought cameras also bought..." can execute within a single partition. Use graph sampling and approximation techniques for analytics queries that don't require exact results—instead of traversing all billions of relationships, sample representative subgraphs and extrapolate. Implement materialized views for common query patterns: if "products frequently bought together" is a common query, precompute and cache these relationship paths rather than traversing the full graph for each query. Deploy distributed graph processing frameworks like Apache Giraph for batch analytics that need to process the entire graph, while using optimized graph databases for real-time queries. Monitor query patterns continuously and adjust partitioning, caching, and indexing strategies based on actual usage.
Challenge: Maintaining Accuracy as Domains Evolve
Knowledge domains evolve continuously—new entities emerge, relationships change, and previously accurate information becomes outdated 45. A relationship map built from 2020 data showing "Company A employs Executive B" becomes inaccurate when Executive B changes employers in 2023, yet the system continues returning outdated information, degrading user trust and potentially causing serious errors in decision-making applications.
Solution:
Implement temporal versioning and continuous update mechanisms with automated staleness detection 6. Tag all relationships with temporal metadata including creation timestamp, last validation timestamp, and optional expiration timestamp. For relationships extracted from dated sources, infer temporal validity: a relationship extracted from a 2020 annual report is valid for 2020 but may not be current. Implement continuous monitoring that re-extracts relationships from updated sources: when a new annual report is published, re-extract employment relationships and compare with existing relationships, updating changed relationships and marking contradictions for review. Deploy change detection algorithms that identify when relationship patterns shift: if multiple recent sources contradict a long-standing relationship, flag it for validation. For critical applications, implement active validation where the system periodically queries authoritative sources to verify key relationships—for example, checking that executive employment relationships match current company websites. Expose temporal queries that allow users to specify "as of" dates: "show me Company A's executive team as of December 2020" retrieves the historical relationship state. Implement relationship confidence decay where confidence scores decrease over time without revalidation, ensuring that old, unverified relationships are eventually flagged for review or excluded from high-confidence queries.
Challenge: Balancing Precision and Recall in Relationship Extraction
Automated relationship extraction faces a fundamental tradeoff between precision (avoiding false positive relationships) and recall (capturing all true relationships) 16. Tuning extraction models for high precision reduces false positives but misses many valid relationships, fragmenting the knowledge graph. Tuning for high recall captures more relationships but introduces noise that degrades downstream applications. In a medical knowledge graph, high precision is critical to avoid suggesting incorrect drug-disease relationships that could harm patients, while high recall is important to avoid missing potentially beneficial treatments.
Solution:
Implement confidence-stratified relationship tiers with application-specific thresholds 7. Configure extraction models to output all candidate relationships with confidence scores, then create multiple relationship tiers: high-confidence relationships (e.g., >0.90) are used directly in production applications, medium-confidence relationships (0.70-0.90) are flagged for expert review before use, and low-confidence relationships (<0.70) are stored for potential future validation but not used in applications. For the medical knowledge graph example, only high-confidence drug-disease relationships would be used in clinical decision support, while medium-confidence relationships would be reviewed by pharmacologists. Implement ensemble methods that combine multiple extraction approaches: if both a rule-based extractor and a neural extractor independently identify the same relationship, confidence increases. Use active learning to prioritize validation efforts: present experts with the medium-confidence relationships that, if validated, would most improve downstream application performance (e.g., relationships connecting previously disconnected graph regions). Create domain-specific precision/recall targets based on application requirements: a fraud detection application might tolerate lower precision (more false positives) to ensure high recall (catching all potential fraud), while a medical application requires high precision even at the cost of recall. Monitor application-level metrics (user satisfaction, task completion rates, error rates) rather than just extraction metrics, and adjust confidence thresholds based on real-world impact.
Challenge: Explaining Relationship-Based Recommendations and Decisions
While relationship mapping enables sophisticated recommendations and decisions, the multi-hop reasoning paths that produce these outputs are often opaque to users, reducing trust and making it difficult to identify errors 48. When a system recommends a particular product or flags a transaction as potentially fraudulent based on traversing a complex path through a knowledge graph, users need to understand the reasoning to trust and act on the recommendation.
Solution:
Implement relationship path visualization and natural language explanation generation 9. When the system makes a recommendation or decision based on graph traversal, capture the specific relationship path that led to the output. For a product recommendation, this might be: User → purchased → Product A → frequently_bought_with → Product B → similar_to → Product C (recommended). Generate natural language explanations from these paths: "We recommend Product C because customers who bought Product A (which you purchased) frequently also bought Product B, and Product C is similar to Product B." Implement interactive visualizations that show the relationship path graphically, allowing users to explore the reasoning. Provide path confidence scores that aggregate the confidence of individual relationships along the path, helping users assess recommendation reliability. For complex multi-hop paths, implement path summarization that highlights the most important relationships while hiding less relevant details. In regulated domains, maintain audit trails of relationship paths used in decisions, enabling post-hoc review and compliance verification. Allow users to provide feedback on explanations ("this recommendation doesn't make sense because..."), and use this feedback to improve relationship extraction and path ranking algorithms.
References
- Rosso, P., Yang, D., & Cudré-Mauroux, P. (2020). Beyond Triplets: Hyper-Relational Knowledge Graph Embedding for Link Prediction. https://arxiv.org/abs/2003.02320
- Bordes, A., Usunier, N., Garcia-Duran, A., Weston, J., & Yakhnenko, O. (2015). Translating Embeddings for Modeling Multi-relational Data. https://arxiv.org/abs/1503.00759
- Dong, X., et al. (2020). Knowledge Graph Construction from Text. https://research.google/pubs/pub48341/
- Xiong, W., Yu, M., Chang, S., Guo, X., & Wang, W. Y. (2021). Improving Question Answering over Incomplete Knowledge Graphs with Multi-Hop Reasoning. https://arxiv.org/abs/2107.07842
- Ji, S., Pan, S., Cambria, E., Marttinen, P., & Yu, P. S. (2020). A Survey on Knowledge Graphs: Representation, Acquisition and Applications. https://www.sciencedirect.com/science/article/pii/S0950705120307516
- Zhang, Y., Zhong, V., Chen, D., Angeli, G., & Manning, C. D. (2019). Position-aware Attention and Supervised Data Improve Slot Filling. https://arxiv.org/abs/1902.10197
- Abu-Salih, B. (2021). Domain-Specific Knowledge Graphs: A Survey. https://ieeexplore.ieee.org/document/9416470
- Yasunaga, M., Ren, H., Bosselut, A., Liang, P., & Leskovec, J. (2020). QA-GNN: Reasoning with Language Models and Knowledge Graphs for Question Answering. https://aclanthology.org/2020.acl-main.521/
- Sun, Z., Deng, Z. H., Nie, J. Y., & Tang, J. (2020). RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space. https://arxiv.org/abs/2010.00904
