Cross-Reference Systems

Cross-Reference Systems in AI Discoverability Architecture represent structured frameworks that establish, maintain, and leverage bidirectional or multidirectional links between related AI artifacts—including machine learning models, training datasets, research papers, API endpoints, and computational workflows—enabling semantic navigation and resource discovery across complex AI ecosystems 12. These systems function as the connective tissue that enables intelligent agents and human users to traverse distributed AI resources, discovering relevant assets through relationship mapping and contextual linking. In an era where AI systems are increasingly distributed, heterogeneous, and specialized, cross-reference systems have become essential infrastructure for preventing information silos, enhancing findability and accessibility, and enabling effective knowledge discovery across organizational and technical boundaries 36.

Overview

The emergence of Cross-Reference Systems in AI Discoverability Architecture stems from the exponential growth and fragmentation of AI resources across the research and development landscape. As machine learning models proliferated and datasets expanded beyond centralized repositories, practitioners faced increasing difficulty locating relevant resources, understanding model lineage, and avoiding redundant development efforts 15. The fundamental challenge these systems address is the discoverability crisis: without structured reference networks, valuable AI assets remain isolated in organizational silos, research papers reference resources without persistent identifiers, and teams unknowingly duplicate work already completed elsewhere 23.

The practice has evolved significantly from early academic citation networks to sophisticated graph-based knowledge systems. Initial approaches borrowed from digital library science, implementing simple metadata catalogs and keyword-based search 1. However, the dynamic nature of AI resources—where models are continuously retrained, datasets updated, and APIs versioned—demanded more sophisticated solutions. Modern cross-reference systems leverage semantic web technologies, graph neural networks, and embedding-based similarity detection to identify both explicit and latent relationships between resources 67. The introduction of standardized metadata frameworks like Model Cards and Data Sheets for Datasets has further accelerated adoption, providing common vocabularies for describing AI artifacts and their relationships 3.

Key Concepts

Entity Resolution

Entity resolution is the process of identifying when different references, identifiers, or descriptions point to the same underlying AI resource 1. This concept addresses the challenge that a single dataset or model may be referenced using different names, versions, or identifiers across various platforms and publications. For example, the ImageNet dataset might be referenced as "ImageNet-1K," "ILSVRC2012," or through various DOIs and URLs across different research papers and model repositories. An effective cross-reference system must recognize these variants as referring to the same canonical resource, enabling accurate relationship mapping and preventing fragmentation of the reference network.

Example: A pharmaceutical company developing drug discovery models encounters references to a protein structure database across multiple sources: their internal documentation calls it "PDB-2023," a research paper references "Protein Data Bank (Release 2023-Q1)," and a partner organization's API documentation uses the identifier "rcsb.org/pdb/v2023.1." The entity resolution component of their cross-reference system analyzes metadata signatures, canonical identifiers, and content hashes to determine these all reference the same resource, consolidating 47 separate internal references into a single canonical entry with known aliases.

Link Typing

Link typing categorizes the semantic relationships between AI resources, moving beyond simple "related-to" connections to capture specific relationship semantics such as "derived-from," "trained-on," "evaluated-on," "compatible-with," or "replaces" 36. These typed relationships enable more sophisticated discovery queries and reasoning about resource dependencies. The relationship ontology defines the vocabulary and semantics of these connections, including hierarchical relationships (parent-child, whole-part), dependency relationships (requires, extends, replaces), and associative relationships (similar-to, related-to).

Example: A computer vision team at an autonomous vehicle company maintains a cross-reference system tracking their perception models. When they register a new pedestrian detection model, they declare typed relationships: "fine-tuned-from" pointing to a base YOLO architecture, "trained-on" linking to their proprietary urban driving dataset, "evaluated-on" referencing the KITTI benchmark, "requires" indicating a specific CUDA version, and "compatible-with" noting integration with their existing sensor fusion pipeline. When a security audit later requires identifying all models trained on data collected in European cities, the typed "trained-on" relationships enable precise queries that would be impossible with untyped generic links.

Reference Integrity

Reference integrity ensures that links between AI resources remain valid and meaningful as resources evolve, are deprecated, or change access permissions 24. This concept addresses the "link rot" problem where references become broken over time, undermining the utility of the cross-reference network. Validation and integrity mechanisms continuously verify reference validity through automated testing and health-check protocols.

Example: A healthcare AI consortium maintains cross-references between diagnostic models and clinical trial datasets. When a hospital updates its patient privacy policies and restricts access to a referenced dataset, the integrity monitoring system detects the access change within 24 hours through automated health checks. It immediately flags 12 dependent model references, notifies affected research teams, updates access metadata to reflect the new restrictions, and suggests three alternative publicly-available datasets with similar characteristics. This prevents researchers from wasting time attempting to access unavailable resources and maintains the accuracy of the reference network.

Provenance Tracking

Provenance tracking maintains comprehensive records of the origin, lineage, and transformation history of AI resources and their references 34. This concept extends beyond simple version control to capture the complete genealogy of models and datasets, including training procedures, data preprocessing steps, and fine-tuning chains. Provenance information becomes critical for reproducibility, compliance, and understanding how properties (including biases or capabilities) propagate through derivative resources.

Example: A financial services firm discovers potential bias in a credit scoring model deployed across 15 regional branches. Their cross-reference system's provenance tracking reveals the model was fine-tuned from a base model trained on demographic data from 2018-2020, which itself derived from an earlier model developed by an acquired company. By traversing the provenance chain, they identify 23 related models sharing the same problematic training data, trace the bias to specific preprocessing decisions made four years earlier, and systematically audit all affected systems. Without provenance tracking, identifying the scope of impact would have required months of manual investigation.

Federated Discovery

Federated discovery enables cross-reference systems to span multiple autonomous registries that synchronize through standardized protocols, allowing organizations to maintain local control while participating in broader discovery ecosystems 27. This approach addresses the challenge that AI resources are inherently distributed across different platforms, cloud providers, and organizational boundaries, making centralized cataloging impractical or undesirable.

Example: A consortium of European research institutions implements federated discovery for climate modeling resources. Each institution maintains its own local registry of models, datasets, and computational resources, but these registries synchronize metadata through a standardized protocol based on schema.org vocabularies. A researcher at the University of Copenhagen searching for ocean temperature datasets receives results not only from their local registry but also from partner institutions in Norway, Germany, and France. The federated system respects each institution's access policies—some resources appear with full metadata while others show only existence and contact information—while providing unified discovery across organizational boundaries.

Graph Traversal and Multi-Hop Reasoning

Graph traversal and multi-hop reasoning enable discovery of non-obvious relationships by exploring indirect connections in the reference network 6. Rather than limiting searches to directly linked resources, these techniques follow chains of relationships to surface relevant resources that may be several steps removed from the initial query. This capability is particularly valuable for discovering alternative resources, understanding complex dependencies, and identifying potential integration opportunities.

Example: A robotics startup searches their cross-reference system for datasets suitable for training a manipulation model for warehouse automation. Their direct search finds three relevant datasets, but the system's multi-hop reasoning explores further: it identifies that one of these datasets was used to evaluate a grasping model, follows references to discover what other datasets were used for training that model, then examines what other models were trained on those datasets, ultimately surfacing a specialized dataset for industrial object manipulation that shares no direct keywords with the original query but proves highly relevant. This discovery, made through three-hop traversal, saves the team four months of data collection effort.

Embedding-Based Similarity

Embedding-based similarity represents AI resources in continuous vector spaces where semantic similarity corresponds to geometric proximity, enabling fuzzy matching and recommendations beyond explicit references 68. This approach uses neural embeddings to capture latent relationships that may not be explicitly declared, complementing traditional graph-based references with learned similarity measures.

Example: A content moderation platform implements embedding-based similarity in their cross-reference system by encoding model descriptions, capabilities, and performance characteristics into 768-dimensional vectors using a specialized language model. When a new policy requires detecting synthetic media, the system recommends relevant detection models not through keyword matching but by finding models whose embedding vectors are geometrically close to the query embedding. This surfaces a video forensics model originally developed for different purposes but whose capability profile—captured in its embedding—indicates strong relevance, despite sharing minimal terminology with the query.

Applications in AI Development and Governance

Cross-Reference Systems find critical applications across the AI development lifecycle, from initial research through production deployment and ongoing governance. In research and development phases, these systems reduce redundant effort by helping practitioners discover existing models and datasets that meet their requirements 15. Organizations with mature cross-reference systems report 30-40% reductions in duplicate model development efforts, as teams can quickly identify whether suitable resources already exist before initiating new projects. For instance, a multinational technology company's internal AI resource catalog, built on cross-reference principles, enables their 200+ machine learning teams to discover and reuse models across divisions, preventing an estimated $15 million annually in redundant development costs.

In model governance and compliance contexts, cross-reference systems enable tracking of model lineage and identifying which production systems depend on specific training datasets or base models 34. When data privacy issues or bias concerns emerge, organizations can rapidly identify all affected downstream resources through reference traversal. A European bank, facing GDPR compliance requirements, used their cross-reference system to trace all models trained on customer data from a specific region, identifying 34 production models and 67 experimental variants that required review—a task that would have taken weeks of manual investigation. This capability has become critical for regulatory compliance under frameworks like the EU AI Act, where demonstrating model provenance and data lineage is increasingly mandatory.

Reproducibility and scientific rigor in AI research depend heavily on robust cross-referencing 5. When research papers reference specific model versions, datasets, and evaluation protocols through persistent identifiers, other researchers can precisely replicate experiments. The absence of effective cross-reference systems contributes to the reproducibility crisis in machine learning research, where studies report that fewer than 30% of published results can be independently verified. Leading AI research institutions now mandate that publications include persistent references to all computational artifacts, with cross-reference systems providing the infrastructure to maintain these links over time.

In collaborative development workflows, cross-reference systems enable distributed teams to maintain coherent understanding of complex AI systems composed of multiple interacting models and data pipelines 7. A global automotive manufacturer developing autonomous driving systems uses cross-reference infrastructure to coordinate work across teams in Germany, the United States, and Japan. Their system tracks dependencies between perception models, sensor calibration datasets, simulation environments, and safety validation protocols, ensuring that updates in one component trigger appropriate reviews and testing in dependent systems. This reference network serves as a shared mental model, reducing coordination overhead and preventing integration conflicts that previously caused costly delays.

Best Practices

Implement Content-Addressable Identifiers for Immutable Resources

AI resources should be assigned persistent identifiers that remain stable even as organizational infrastructure changes 23. For immutable resources like specific dataset versions or trained model snapshots, content-addressable identifiers (similar to Git's SHA hashes) provide cryptographic guarantees of identity. For evolving resources, semantic versioning distinguishes between minor updates that preserve identifiers and major changes requiring new identifiers.

Rationale: Persistent identification prevents reference rot and enables long-term reproducibility. When identifiers are tied to organizational infrastructure (like internal server names or temporary URLs), references break when infrastructure changes. Content-addressable schemes ensure that the identifier itself encodes resource identity, remaining valid regardless of storage location.

Implementation Example: A genomics research institute implements content-addressable identifiers for their DNA sequence datasets using SHA-256 hashes of dataset contents. Each dataset receives an identifier like dataset:sha256:a3f5b9c2... that uniquely identifies that specific version. When they migrate from on-premises storage to cloud infrastructure, all existing references remain valid because the identifiers are content-based rather than location-based. Their cross-reference system maintains a resolution service that maps these identifiers to current storage locations, allowing seamless access despite infrastructure changes.

Start with Minimal Viable Ontologies and Expand Based on Demonstrated Need

Organizations should begin with simple relationship types and expand their ontology incrementally as specific use cases demonstrate value 6. Over-engineering ontologies with excessive relationship types creates confusion and reduces adoption, while overly simplistic schemes limit utility.

Rationale: Complex ontologies require significant maintenance effort and user training. Many organizations create elaborate relationship taxonomies that users find confusing or ignore, resulting in inconsistent application. Starting minimal and expanding based on actual usage patterns ensures the ontology remains practical and aligned with real needs.

Implementation Example: A pharmaceutical company initially implements only three relationship types in their drug discovery cross-reference system: "derived-from" for model lineage, "trained-on" for data dependencies, and "evaluated-on" for benchmark relationships. After six months of usage data, they identify that users frequently add free-text notes about molecular similarity between datasets. This demonstrates clear need, so they add a formal "similar-to" relationship type with structured similarity metrics. Over two years, they expand to 12 relationship types, each added only after demonstrating value through usage patterns, resulting in 94% consistent application compared to 31% consistency in their previous system with 45 pre-defined relationship types.

Implement Automated Reference Validation with Graceful Degradation

Cross-reference systems should continuously monitor reference validity through automated health checks while providing graceful degradation when resources become unavailable 24. Rather than simply marking references as broken, systems should capture context about failures and suggest alternatives.

Rationale: AI resources change frequently—models are retrained, datasets updated, APIs versioned, and access permissions modified. Without continuous validation, reference networks rapidly accumulate broken links that undermine utility. Graceful degradation maintains partial functionality even when specific resources are unavailable.

Implementation Example: A computer vision platform implements daily automated validation that attempts to access each referenced resource and verify its metadata signature. When a referenced dataset becomes unavailable, rather than simply marking it as broken, the system: (1) captures the failure context (access denied vs. resource not found vs. network error), (2) checks if newer versions exist at the same location, (3) searches for resources with similar metadata signatures, (4) notifies resource owners and dependent users, and (5) maintains cached metadata so users can still discover what the resource was even if they can't currently access it. This approach reduced user-reported broken reference complaints by 78% compared to their previous binary available/unavailable system.

Provide Both Human and Machine Interfaces

Effective cross-reference systems must serve both human practitioners exploring resources interactively and automated agents integrating discovery into computational workflows 17. This requires parallel investment in intuitive web interfaces and well-documented APIs with client libraries in multiple programming languages.

Rationale: Human users need visual exploration, contextual information, and serendipitous discovery capabilities, while automated systems require programmatic access, batch operations, and webhook notifications. Systems optimized for only one audience fail to achieve full value.

Implementation Example: A national AI research infrastructure provides a web interface with graph visualization showing model relationships, faceted search with filters for model architecture and performance metrics, and curated collections for common use cases. Simultaneously, they offer a REST API with comprehensive OpenAPI documentation, Python and R client libraries, SPARQL endpoints for complex graph queries, and webhook subscriptions for change notifications. Usage analytics show 60% of discovery sessions use the web interface while 85% of actual resource retrievals occur through API calls, demonstrating that both interfaces serve essential but different roles in the discovery workflow.

Implementation Considerations

Tool and Format Choices

Selecting appropriate technologies for cross-reference systems requires balancing interoperability, performance, and organizational expertise 16. Organizations must choose between native graph databases (Neo4j, Amazon Neptune) optimized for traversal queries, triple stores (Apache Jena, Virtuoso) providing semantic web standards compliance, or hybrid approaches combining relational databases with graph query layers. Format choices include RDF for maximum interoperability with semantic web ecosystems, property graphs for more intuitive modeling, or custom schemas optimized for specific use cases.

Example: A healthcare AI consortium evaluates technology options for their cross-reference system. They select Amazon Neptune with property graph model because: (1) their team has existing AWS expertise, (2) property graphs provide more intuitive modeling for their use cases than RDF triples, (3) Neptune's managed service reduces operational overhead, and (4) Gremlin query language offers sufficient expressiveness for their traversal needs. However, they implement an RDF export layer using schema.org vocabularies to enable interoperability with external research databases, demonstrating that technology choices need not be exclusive.

Audience-Specific Customization

Cross-reference systems serve diverse audiences with different needs—data scientists seeking training datasets, compliance officers tracking model lineage, executives monitoring resource utilization, and automated systems orchestrating workflows 34. Effective implementations provide customized views and query capabilities tailored to each audience while maintaining a unified underlying reference network.

Example: A financial services firm implements role-based views in their AI resource cross-reference system. Data scientists see detailed technical metadata, performance benchmarks, and code examples. Compliance officers see audit trails, data lineage, and regulatory classification. Executives see aggregate metrics, cost information, and utilization trends. Model deployment systems access machine-readable dependency graphs and compatibility matrices. All views query the same underlying reference network but present information appropriate to each audience's needs and expertise level, increasing adoption across all user groups.

Organizational Maturity and Context

Implementation approaches must align with organizational AI maturity and governance structures 49. Organizations with nascent AI practices benefit from lightweight, low-friction systems that encourage voluntary participation, while mature organizations with established governance can implement more comprehensive mandatory registration and validation requirements.

Example: A retail company in early AI adoption stages implements a voluntary cross-reference system where teams can optionally register models and datasets, with minimal required metadata and no enforcement mechanisms. As their AI practice matures over three years and governance requirements increase, they gradually introduce mandatory registration for production models, expand required metadata fields, implement automated compliance checks, and integrate the cross-reference system with their model deployment pipeline. This phased approach achieves 89% registration compliance compared to 34% in a peer organization that attempted comprehensive mandatory requirements from inception, which faced resistance and workarounds.

Privacy and Access Control Integration

Cross-reference systems spanning public and proprietary resources must carefully balance discoverability with access control 29. Systems should implement reference-level permissions where the existence and metadata of a relationship may be visible to facilitate discovery while the referenced resource itself remains access-controlled.

Example: A multi-national corporation implements tiered metadata visibility in their cross-reference system. Public tier metadata (model type, general capabilities, publication references) is visible to all employees. Restricted tier metadata (specific performance metrics, training data characteristics) requires project membership. Confidential tier metadata (proprietary datasets, competitive advantage details) requires explicit approval. The reference graph itself is visible to all users, enabling discovery of relevant resources, but detailed metadata and access to actual resources respects existing access controls. This approach increased cross-divisional resource discovery by 340% while maintaining security requirements, as teams could discover relevant resources existed and request access rather than remaining unaware of their existence.

Common Challenges and Solutions

Challenge: Identifier Persistence Across Organizational Change

AI resources evolve rapidly, with models being retrained, datasets updated, and APIs versioned, while organizational infrastructure changes through migrations, mergers, and technology transitions 2. Identifiers tied to specific infrastructure or organizational structures break when these change, undermining the long-term value of reference networks. A common scenario involves references using internal server names or department-specific URLs that become invalid when infrastructure is modernized or organizational units are restructured.

Solution:

Implement a multi-layered identifier strategy combining content-addressable identifiers for immutable resources with persistent identifier services for evolving resources 3. Content-addressable schemes (like SHA-256 hashes of model weights or dataset contents) provide cryptographic guarantees of identity independent of storage location. For evolving resources, implement a persistent identifier service (similar to DOI resolution) that maintains mappings between stable identifiers and current locations, updating mappings as infrastructure changes. A technology company implementing this approach assigns identifiers like model:uuid:a7f3c9d2-4b8e-11ed-bdc3-0242ac120002 for evolving models and model:sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 for specific immutable versions, with a resolution service maintaining current access endpoints. When they migrated from on-premises to cloud infrastructure, all existing references remained valid because identifiers were infrastructure-independent, with only the resolution service mappings requiring updates.

Challenge: Scale and Performance with Growing Reference Networks

As cross-reference networks grow to millions of nodes and billions of edges, query performance degrades and storage costs escalate 6. A research consortium experienced this when their reference network grew from 50,000 resources with 200,000 relationships to 5 million resources with 80 million relationships over three years, causing query response times to increase from milliseconds to minutes and making the system unusable for interactive discovery.

Solution:

Implement hierarchical indexing and materialized views for common query patterns, combined with graph partitioning strategies for distributed storage 16. Hierarchical indexing creates multiple resolution levels—a fast in-memory index for frequently accessed resources, a mid-tier cache for moderately popular items, and full graph storage for comprehensive queries. Materialized views pre-compute common traversal patterns (like "all datasets used to train models in the computer vision domain") and update incrementally as the graph changes. Graph partitioning distributes the reference network across multiple servers based on access patterns and relationship locality. The research consortium implemented this approach by: (1) maintaining a Redis cache of the 100,000 most-accessed resources, (2) pre-computing materialized views for the 50 most common query patterns, refreshed nightly, (3) partitioning their graph database by domain area (computer vision, NLP, robotics, etc.) with cross-partition queries handled through a federation layer. This reduced median query response time from 45 seconds to 120 milliseconds while supporting 10x growth in reference network size.

Challenge: Reference Quality and Spam Prevention

In open ecosystems where anyone can register resources and declare relationships, reference quality degrades through spam, erroneous links, and low-quality metadata 14. A public AI model repository experienced this when automated bots began registering thousands of fake models with spurious relationships to popular resources, attempting to boost visibility through reference manipulation, degrading discovery quality for legitimate users.

Solution:

Implement multi-layered quality assurance combining automated validation, reputation mechanisms, and community moderation 49. Automated validation detects suspicious patterns like circular references, orphaned resources, excessive outbound links, or metadata inconsistencies. Reputation mechanisms weight references based on the authority of the declaring entity, derived from community validation and usage patterns. Community moderation enables users to flag problematic references for review. The model repository implemented this by: (1) automated checks rejecting registrations with more than 20 outbound references or metadata fields containing spam keywords, (2) a reputation system where references from verified organizations and frequently-used resources receive higher weight in discovery ranking, (3) community flagging with moderator review for suspicious content, and (4) rate limiting preventing individual accounts from registering more than 10 resources per day. This reduced spam registrations by 94% while maintaining low friction for legitimate users, with 99.2% of legitimate registrations passing automated validation without manual review.

Challenge: Maintaining Reference Integrity During Resource Evolution

AI resources frequently undergo updates—models are retrained with new data, datasets are expanded or corrected, and APIs are versioned—creating challenges for maintaining meaningful references 23. A common problem occurs when a reference to "the ImageNet dataset" becomes ambiguous as multiple versions with different characteristics exist, or when a model is significantly updated but retains the same identifier, breaking assumptions made by dependent systems.

Solution:

Implement comprehensive versioning with explicit deprecation workflows and automated dependency notification 34. Every resource should support versioning that distinguishes between minor updates (bug fixes, documentation improvements) preserving compatibility and major updates requiring new identifiers. Deprecation workflows provide advance notice when resources will be retired, suggest migration paths to alternatives, and maintain read-only access to deprecated resources for reproducibility. Automated dependency notification alerts owners of dependent resources when referenced items are updated or deprecated. A pharmaceutical company implemented this by: (1) requiring semantic versioning (major.minor.patch) for all registered resources, (2) a six-month deprecation notice period during which deprecated resources remain accessible but display warnings, (3) automated emails to owners of dependent resources when dependencies are updated, including change summaries and compatibility assessments, and (4) maintaining archived access to all deprecated resources for reproducibility, even after active support ends. This approach reduced breaking changes affecting production systems by 87% while enabling continuous improvement of resources.

Challenge: Cross-Organizational Discovery and Privacy

Organizations increasingly need to discover resources across organizational boundaries while respecting proprietary information and access controls 29. A healthcare research consortium faced this challenge when member institutions wanted to discover each other's clinical datasets and models without exposing sensitive details about patient populations or proprietary methodologies to competitors or unauthorized parties.

Solution:

Implement federated discovery with tiered metadata visibility and privacy-preserving techniques 79. Federated architectures allow each organization to maintain local control over their resources while participating in broader discovery through standardized protocols. Tiered metadata visibility exposes different information levels based on relationship and authorization—basic existence and contact information may be public, while detailed characteristics require partnership agreements, and full access requires explicit authorization. Privacy-preserving techniques like differential privacy enable aggregate statistics (e.g., "15 institutions have datasets with similar characteristics") without exposing individual details. The healthcare consortium implemented this by: (1) each institution maintaining a local registry with full metadata, (2) a federated discovery layer exposing only basic metadata (resource type, institution, contact) publicly, (3) partnership agreements enabling detailed metadata sharing between specific institutions, (4) differential privacy for aggregate queries across the consortium, and (5) access request workflows connecting interested parties while respecting institutional policies. This enabled discovery of relevant resources across organizational boundaries while maintaining privacy and competitive protections, increasing cross-institutional collaborations by 230%.

References

  1. Chapman, A., et al. (2018). Dataset Search: A Survey. arXiv. https://arxiv.org/abs/1810.03993
  2. Noy, N., et al. (2019). Google Dataset Search. Google Research Publications. https://research.google/pubs/pub46846/
  3. Mitchell, M., et al. (2021). Model Cards for Model Reporting. arXiv. https://arxiv.org/abs/2108.07258
  4. Paleyes, A., et al. (2021). AI Model Governance. IEEE. https://ieeexplore.ieee.org/document/9458899
  5. Pineau, J., et al. (2020). ML Reproducibility. arXiv. https://arxiv.org/abs/2011.03395
  6. Rossi, A., et al. (2021). Knowledge Graph Embeddings. ScienceDirect. https://www.sciencedirect.com/science/article/pii/S0306437921000326
  7. Bommasani, R., et al. (2021). Foundation Models. arXiv. https://arxiv.org/abs/2103.00020
  8. Goh, G., et al. (2021). Multimodal Neurons. Distill. https://distill.pub/2021/multimodal-neurons/
  9. Crisan, A., et al. (2022). Responsible AI Practices. Google Research Publications. https://research.google/pubs/pub49953/