Document Chunking Strategies
Document chunking strategies represent a critical preprocessing methodology in AI discoverability architecture, involving the systematic decomposition of large text documents into smaller, semantically coherent segments for optimal retrieval and processing 1. The primary purpose of chunking is to enable efficient information retrieval in vector databases and retrieval-augmented generation (RAG) systems by creating text segments that balance semantic completeness with computational efficiency 23. This approach matters fundamentally because it directly impacts the quality of semantic search, the accuracy of question-answering systems, and the overall effectiveness of large language model (LLM) applications that depend on external knowledge retrieval 12. As organizations increasingly deploy AI systems that must discover and utilize information from vast document repositories, the strategic implementation of chunking methodologies has become essential for achieving high-precision information retrieval and maintaining contextual coherence in AI-generated responses 38.
Overview
The emergence of document chunking strategies as a distinct discipline within AI architecture stems from the convergence of several technological developments in natural language processing and information retrieval. The fundamental challenge that chunking addresses is the inherent tension between the limited context windows of embedding models and the need to process extensive document collections while preserving semantic meaning 12. Early information retrieval systems relied on keyword matching and document-level indexing, but the advent of dense vector representations and neural embedding models created new opportunities—and requirements—for more granular text segmentation 4.
The practice has evolved significantly from simple fixed-size splitting to sophisticated semantic and hierarchical approaches that respect document structure and topical boundaries 35. Modern chunking strategies emerged in response to the limitations of embedding models, which typically have maximum token limits ranging from 512 to 8,192 tokens depending on the architecture 16. As retrieval-augmented generation systems became prevalent in enterprise applications, the quality of chunking directly influenced the accuracy and relevance of AI-generated responses, elevating chunking from a preprocessing afterthought to a strategic architectural decision 78.
Key Concepts
Chunk Size Parameter
Chunk size refers to the maximum length of each text segment, typically measured in tokens or characters, and represents the most fundamental parameter in any chunking strategy 1. This parameter must be calibrated based on the embedding model's capacity, the nature of source documents, and the specificity required for retrieval tasks 23.
For example, a medical knowledge base implementing a RAG system for clinical decision support might use 512-token chunks when processing treatment protocols. This size ensures that each chunk contains a complete treatment step—including the intervention, dosage, timing, and contraindications—without fragmenting critical safety information across multiple segments. The system uses OpenAI's text-embedding-ada-002 model, which handles up to 8,191 tokens but performs optimally with shorter, focused segments that produce more discriminative embeddings for precise retrieval during physician queries.
Chunk Overlap
Chunk overlap constitutes the degree of content duplication between adjacent chunks, serving as a crucial mechanism for maintaining continuity across chunk boundaries and preventing information loss 15. Overlap percentages typically range from 10-20% of chunk size, ensuring that concepts spanning boundaries remain retrievable 2.
Consider a legal contract analysis system processing a 50-page merger agreement. The system implements 768-token chunks with 20% overlap (approximately 154 tokens). When a critical liability clause spans a natural chunk boundary, the 154-token overlap ensures that the complete clause appears in both adjacent chunks. During retrieval, when an attorney queries about indemnification provisions, the system successfully retrieves the relevant chunk containing the complete clause context, rather than returning a fragment that begins mid-sentence due to an arbitrary split point.
Semantic Boundaries
Semantic boundaries represent natural divisions in content structure—such as paragraphs, sections, and thematic transitions—that should guide chunking decisions rather than arbitrary character counts 36. This concept recognizes that documents possess inherent organizational logic that, when preserved, enhances both retrieval precision and result interpretability 5.
A technical documentation system for enterprise software illustrates this concept. Rather than splitting a troubleshooting guide at a fixed 600-token interval that might separate a problem description from its solution steps, the system analyzes document structure to identify semantic boundaries. It detects heading markers like "Problem: Database Connection Timeout" and "Solution:" as natural split points, creating chunks that contain complete troubleshooting units. This approach ensures that when a developer searches for connection timeout solutions, they retrieve a coherent chunk containing both the diagnostic steps and the complete resolution procedure.
Hierarchical Chunking
Hierarchical chunking creates multi-level representations where parent chunks contain section summaries or overviews while child chunks contain detailed content, forming a tree structure that enables both broad and narrow retrieval 37. This methodology proves particularly effective for long-form technical documentation and academic papers 5.
An academic research database implementing hierarchical chunking processes a 15,000-word neuroscience paper by creating a parent chunk containing the abstract and section headings (approximately 400 tokens), with child chunks for each major section: Introduction (3 chunks of 800 tokens each), Methods (5 chunks), Results (6 chunks), and Discussion (4 chunks). When a researcher queries about "synaptic plasticity mechanisms," the system first searches child chunks for precision, identifying relevant passages in the Results section. However, it returns the parent chunk alongside the specific child chunks, providing the researcher with both the detailed findings and the broader paper context, including the research question and methodology overview.
Metadata Preservation
Metadata preservation ensures that each chunk retains contextual information about its source, including document title, section headers, page numbers, and hierarchical position 28. This metadata enriches retrieval results and enables sophisticated filtering and ranking mechanisms 3.
A corporate knowledge management system processing internal policy documents demonstrates metadata preservation in practice. When chunking the 120-page Employee Handbook, each chunk receives metadata tags including: document_title: "Employee Handbook 2024", section: "Benefits and Compensation", subsection: "Health Insurance Options", page_range: "45-47", and last_updated: "2024-01-15". When an HR representative searches for health insurance information, the system not only retrieves relevant chunks but displays the section hierarchy and page numbers, allowing the representative to quickly verify the information's currency and locate the content in the source document for official reference.
Embedding Similarity Thresholds
Embedding similarity thresholds define the minimum cosine similarity score between consecutive text segments that determines whether they belong to the same semantic unit or represent a topic transition warranting a chunk boundary 35. This concept underpins semantic chunking approaches that group related sentences based on topical coherence 6.
A news aggregation platform implementing semantic chunking processes long-form investigative journalism articles. The system calculates embeddings for each sentence using a sentence-transformer model, then measures cosine similarity between consecutive sentences. When processing an article about climate policy that transitions from discussing carbon emissions (sentences 1-8, with inter-sentence similarity scores of 0.82-0.89) to renewable energy subsidies (sentences 9-15, with similarity scores of 0.85-0.91), the system detects a similarity drop to 0.54 between sentences 8 and 9. This drop below the configured threshold of 0.65 triggers a new chunk boundary, ensuring that carbon emissions content and renewable energy content are indexed separately for more precise topical retrieval.
Applications in Information Retrieval Systems
Customer Support Knowledge Bases
Document chunking strategies find extensive application in customer support systems where rapid, accurate information retrieval directly impacts customer satisfaction and support efficiency 28. These systems typically implement recursive character splitting with domain-specific boundary markers to maintain the integrity of question-answer pairs and troubleshooting procedures 3.
A telecommunications company's support chatbot processes a knowledge base containing 5,000 support articles covering network troubleshooting, billing inquiries, and device setup. The system implements 600-token chunks with 15% overlap, using recursive splitting that prioritizes breaking at FAQ boundaries (identified by question marks followed by answer sections) and procedural step markers (numbered lists, bullet points). When a customer asks, "Why is my internet connection dropping frequently?", the system retrieves a coherent chunk containing the complete diagnostic flowchart—from checking physical connections through router configuration verification—rather than fragmenting the procedure across multiple retrievals that would confuse both the chatbot's response generation and the customer's troubleshooting process.
Legal Document Analysis
Legal document analysis systems employ hierarchical chunking to maintain relationships between contract clauses, statutory provisions, and their parent sections while enabling precise retrieval of specific legal language 57. The complex, nested structure of legal documents—with articles, sections, subsections, and clauses—demands sophisticated chunking that preserves this hierarchy 3.
A contract review platform processing commercial lease agreements implements three-level hierarchical chunking: Level 1 parent chunks contain article summaries (e.g., "Article V: Maintenance and Repairs"), Level 2 chunks contain section-level content (e.g., "Section 5.2: Tenant Responsibilities"), and Level 3 child chunks contain specific clause language with 512-token granularity. When an attorney searches for "HVAC maintenance obligations," the system retrieves the specific Level 3 chunk containing the detailed HVAC clause language, but also provides the Level 2 section context showing all tenant maintenance responsibilities and the Level 1 article overview, enabling the attorney to understand the specific obligation within the broader maintenance framework of the lease.
Medical Knowledge Retrieval
Medical knowledge bases utilize semantic chunking to ensure that symptom descriptions, diagnostic criteria, treatment protocols, and contraindications remain coherent units that support accurate clinical decision-making 16. The high stakes of medical information retrieval demand chunking strategies that prevent dangerous fragmentation of critical safety information 8.
A clinical decision support system processing medical literature and treatment guidelines implements semantic chunking with domain-specific constraints. When processing a cardiology guideline document describing acute myocardial infarction treatment, the system uses embedding similarity to group related content but enforces a rule that contraindication information must never be separated from its associated treatment recommendation. For the section on thrombolytic therapy, the system creates a 680-token chunk containing the treatment indication, dosing protocol, administration timing, and complete contraindication list (including recent surgery, bleeding disorders, and uncontrolled hypertension). This ensures that when a physician queries about thrombolytic therapy, the retrieved chunk always includes the critical safety information alongside the treatment protocol, preventing potentially dangerous partial information retrieval.
Enterprise Search and Knowledge Management
Enterprise search systems apply document chunking to vast repositories of internal documentation, enabling employees to discover relevant information across diverse document types including technical specifications, process documentation, and project reports 27. These systems must handle heterogeneous document structures while maintaining consistent retrieval quality 38.
A multinational corporation's internal search platform processes over 500,000 documents spanning engineering specifications, HR policies, financial reports, and project documentation. The system implements adaptive chunking that selects strategies based on document type: technical specifications use 800-token chunks with structural boundary detection (respecting section headers and specification tables), HR policies use 512-token chunks with 20% overlap to ensure policy provisions remain complete, and project reports use semantic chunking to group related project phases. When an engineer searches for "thermal management requirements for data center equipment," the system retrieves precisely targeted chunks from engineering specifications that contain complete thermal specification tables and testing protocols, rather than fragmenting the technical requirements across multiple incomplete segments.
Best Practices
Conduct Systematic Chunk Size Optimization
Empirical testing to determine optimal chunk size for specific use cases represents a fundamental best practice, as optimal sizes vary significantly across domains and retrieval objectives 12. The rationale stems from research demonstrating that chunk size directly affects retrieval precision, with the ideal balance between granularity and context preservation depending on document characteristics and query patterns 3.
Implementation involves creating a test set of representative queries and documents, then systematically evaluating retrieval metrics (precision, recall, Mean Reciprocal Rank) across chunk sizes ranging from 256 to 1,536 tokens in 128-token increments. For example, a financial services firm implementing a RAG system for investment research discovered through testing that 640-token chunks optimized retrieval for their use case: smaller 384-token chunks fragmented complex financial analyses across multiple segments, reducing answer coherence, while larger 1,024-token chunks introduced excessive noise from tangential content, degrading retrieval precision. The 640-token configuration captured complete analytical sections (thesis, supporting data, conclusion) while maintaining focused semantic signals.
Implement Appropriate Overlap to Prevent Boundary Information Loss
Configuring chunk overlap between 10-20% of chunk size prevents critical information loss at boundaries while managing storage and computational costs 15. Research demonstrates that overlap ensures concepts spanning chunk boundaries remain retrievable, particularly important for documents where key information may appear at arbitrary positions 2.
A pharmaceutical company's drug interaction database implements 512-token chunks with 20% overlap (102 tokens) when processing clinical pharmacology literature. This configuration proved essential when processing a paper describing a critical drug interaction between anticoagulants and NSAIDs. The key interaction mechanism explanation spanned a natural paragraph boundary that would have fallen at a chunk split point. The 102-token overlap ensured that both adjacent chunks contained the complete mechanism description, guaranteeing that queries about either drug class would retrieve the full interaction context. Testing revealed that reducing overlap to 5% resulted in a 23% increase in incomplete information retrieval for boundary-spanning content.
Preserve Hierarchical Context Through Metadata
Maintaining comprehensive metadata that captures document hierarchy, source references, and contextual information enables more sophisticated retrieval and improves result interpretability 23. The rationale recognizes that chunks divorced from their source context lose significant meaning, and metadata provides the scaffolding for reconstructing this context during retrieval 8.
An aerospace engineering documentation system implements rich metadata preservation when chunking technical manuals. Each chunk receives metadata including: manual_title, chapter_number, section_hierarchy (e.g., "3.2.4 Hydraulic System Pressure Testing"), page_numbers, document_version, last_revision_date, and safety_critical_flag. When a maintenance technician queries about hydraulic pressure testing procedures, the system retrieves relevant chunks and displays them with full hierarchical context: "Manual: Boeing 737 Maintenance > Chapter 3: Hydraulic Systems > Section 3.2: Testing Procedures > 3.2.4: Pressure Testing (Pages 145-147, Rev. 2024-01, Safety Critical)." This metadata enables technicians to verify they're referencing current, authoritative procedures and understand the retrieved information's position within the broader maintenance framework.
Validate Chunking Quality Through Retrieval Testing
Implementing systematic retrieval testing with representative queries ensures that chunking strategies achieve desired retrieval outcomes rather than optimizing theoretical metrics divorced from actual use cases 37. This practice recognizes that chunking effectiveness can only be truly evaluated through its impact on end-to-end retrieval performance 8.
A legal research platform validates its hierarchical chunking strategy by maintaining a test set of 500 representative legal queries with human-annotated relevant passages. After implementing a new chunking approach that creates parent chunks for case summaries and child chunks for specific legal holdings, the team runs the complete test set and measures whether the top-3 retrieved chunks contain the human-annotated relevant passages. Initial testing revealed that 78% of queries successfully retrieved relevant content, but 22% failed because important procedural context was separated from substantive holdings. The team refined the chunking logic to keep procedural context with related holdings, improving retrieval success to 91%, demonstrating how systematic testing drives iterative chunking optimization.
Implementation Considerations
Tool and Framework Selection
Selecting appropriate chunking tools and frameworks depends on document complexity, processing volume, and integration requirements with existing infrastructure 23. Organizations must evaluate trade-offs between general-purpose frameworks like LangChain and LlamaIndex versus custom implementations using lower-level libraries like spaCy or NLTK 5.
A media company processing diverse content types—from short news articles to long-form investigative reports and multimedia transcripts—initially implemented LangChain's RecursiveCharacterTextSplitter for its simplicity and rapid deployment. However, as requirements evolved to include semantic chunking for investigative pieces and custom boundary detection for interview transcripts (splitting at speaker changes), the team developed a hybrid approach: LangChain for standard news articles, a custom semantic chunker using sentence-transformers for long-form content, and a specialized transcript processor using spaCy for speaker identification. This multi-tool strategy balanced development velocity with chunking quality across diverse content types.
Embedding Model Compatibility
Chunking strategies must align with the capabilities and limitations of the selected embedding model, including maximum token limits, optimal input length for embedding quality, and model-specific tokenization 16. Different embedding models exhibit varying performance characteristics across input lengths, requiring chunking calibration 2.
A multilingual customer support system using the paraphrase-multilingual-mpnet-base-v2 embedding model (maximum 512 tokens) initially implemented 500-token chunks, assuming this would maximize information density while staying within limits. However, testing revealed that this model produces higher-quality embeddings for 256-384 token inputs, where the semantic signal-to-noise ratio is optimal. The team reconfigured to 350-token chunks with 15% overlap, resulting in a 12% improvement in retrieval precision. Additionally, they discovered that the model's tokenization produced different token counts for the same character length across languages (English, Spanish, Mandarin), requiring language-specific chunk size calibration to maintain consistent semantic granularity.
Domain-Specific Structural Conventions
Different document domains possess distinct structural conventions that should inform chunking strategies—legal documents have articles and clauses, scientific papers have IMRaD structure (Introduction, Methods, Results, Discussion), and technical manuals have hierarchical procedure sections 35. Recognizing and leveraging these conventions improves chunking quality 7.
A scientific literature database processing biomedical research papers implements domain-aware chunking that recognizes IMRaD structure. The system uses different strategies for each section: Methods sections receive larger 900-token chunks because experimental procedures require complete protocol descriptions; Results sections use 600-token chunks aligned with individual experiments or figure discussions; Discussion sections employ semantic chunking to group related interpretive arguments. This domain-specific approach outperformed generic fixed-size chunking by 34% in retrieval precision tests, as it preserved the logical units that researchers actually seek when querying the literature (complete experimental methods, specific result sets, coherent interpretive arguments).
Scalability and Performance Requirements
Implementation must consider processing throughput, storage costs, and query latency requirements, as chunking decisions directly impact system scalability 28. Smaller chunks increase vector database size and potentially query latency, while larger chunks reduce these costs but may compromise retrieval precision 3.
An e-commerce company's product information retrieval system processes 10 million product descriptions, technical specifications, and user manuals. Initial implementation with 256-token chunks created 45 million vectors, resulting in a vector database requiring 180GB storage and 250ms average query latency. Analysis revealed that many product descriptions were under 400 tokens, making fine-grained chunking unnecessary. The team implemented adaptive chunking: products with descriptions under 500 tokens are indexed as single chunks, while longer technical manuals use 512-token hierarchical chunks. This reduced the vector count to 18 million, decreased storage to 72GB, and improved query latency to 95ms while maintaining retrieval quality, demonstrating how scalability considerations should inform chunking granularity decisions.
Common Challenges and Solutions
Challenge: Fragmenting Critical Information Across Chunk Boundaries
One of the most significant challenges in document chunking involves the risk of splitting semantically related information—such as a problem description and its solution, a claim and its supporting evidence, or a procedure and its safety warnings—across multiple chunks, degrading retrieval quality and potentially creating dangerous incomplete information retrieval 13. This challenge intensifies with fixed-size chunking approaches that ignore semantic and structural boundaries 5.
Solution:
Implement semantic boundary detection combined with strategic overlap configuration to maintain information coherence. For structured documents, parse explicit markers (headings, list structures, question-answer pairs) and enforce rules that keep related elements together. For unstructured text, employ embedding similarity analysis to detect topic transitions and avoid splitting within high-coherence segments 36.
A pharmaceutical regulatory documentation system illustrates this solution. When processing drug safety monographs, the system implements a rule-based approach that identifies structural patterns: adverse event descriptions are always followed by frequency data and severity classifications. The chunking logic detects these patterns using regular expressions and heading analysis, creating chunks that contain complete adverse event profiles (event description + frequency + severity + management recommendations) even if this results in variable chunk sizes ranging from 400-800 tokens. Additionally, the system implements 25% overlap specifically for safety-critical sections, ensuring that any boundary-spanning safety information appears completely in at least one chunk. This approach eliminated the 15% incomplete safety information retrieval rate observed with fixed-size chunking.
Challenge: Handling Diverse Document Formats and Structures
Organizations typically maintain document repositories containing heterogeneous formats—PDFs with complex layouts, HTML with embedded navigation, Word documents with tables and images, and scanned documents requiring OCR—each presenting distinct parsing and chunking challenges 28. Inconsistent formatting, extraction artifacts, and structural ambiguity complicate the creation of clean, semantically coherent chunks 3.
Solution:
Develop a multi-stage preprocessing pipeline that normalizes diverse formats into a standardized intermediate representation while preserving structural information, followed by format-specific chunking logic that adapts to document characteristics 57. Implement robust error handling and quality validation to detect and remediate extraction artifacts.
A legal services firm processing court documents, contracts, and case law from multiple sources implemented a three-stage pipeline: (1) Format detection and specialized extraction using Apache Tika for PDFs, BeautifulSoup for HTML, and python-docx for Word documents, with OCR fallback for scanned documents; (2) Structural normalization that converts all documents to a common JSON schema preserving headings, paragraphs, lists, and tables with hierarchy metadata; (3) Adaptive chunking that selects strategies based on detected document type (contracts use hierarchical chunking respecting article/section structure, court opinions use semantic chunking for legal arguments, case law summaries use fixed-size chunking with overlap). The system includes validation that flags chunks with excessive formatting artifacts, incomplete sentences, or anomalous length distributions for manual review. This approach reduced chunking errors from 23% to 4% across the heterogeneous document collection.
Challenge: Optimizing Chunk Size for Diverse Query Types
Different query types and information needs require different levels of granularity—specific factual queries benefit from fine-grained chunks that enable precise retrieval, while complex analytical queries require larger chunks that provide sufficient context for comprehensive answers 12. A single chunk size cannot optimally serve all query patterns, yet maintaining multiple chunking strategies increases system complexity 3.
Solution:
Implement hierarchical or multi-resolution chunking that creates chunks at multiple granularity levels, enabling the retrieval system to select appropriate resolution based on query characteristics 57. Alternatively, use query analysis to dynamically adjust retrieval parameters or employ a two-stage retrieval process that first identifies relevant document sections, then retrieves fine-grained chunks within those sections.
A business intelligence platform supporting both quick fact lookups ("What was Q3 revenue?") and complex analytical queries ("Analyze factors contributing to regional sales variance") implements three-tier chunking for financial reports: Tier 1 creates document-level summaries (executive summary, key metrics); Tier 2 creates section-level chunks (600-800 tokens covering complete analytical sections); Tier 3 creates fine-grained chunks (300-400 tokens for specific data points and metrics). The query processing pipeline analyzes query complexity using a classifier trained on historical queries: simple factual queries retrieve from Tier 3 for precision, analytical queries retrieve from Tier 2 for context, and exploratory queries start with Tier 1 then drill down to relevant Tier 2 sections. This multi-resolution approach improved user satisfaction scores by 28% compared to single-resolution chunking, as it matched retrieval granularity to information needs.
Challenge: Maintaining Chunking Quality at Scale
As document collections grow to millions of documents, maintaining consistent chunking quality becomes challenging—processing pipelines must handle edge cases, document quality varies widely, and manual quality review becomes impractical 28. Performance requirements may pressure teams to adopt simpler chunking strategies that sacrifice quality for throughput 3.
Solution:
Implement automated quality monitoring that samples chunked output and measures quality metrics including chunk coherence scores, boundary quality (percentage of chunks ending mid-sentence), metadata completeness, and retrieval performance on test queries 67. Use these metrics to identify problematic document types or processing failures, and implement continuous improvement processes that refine chunking logic based on quality monitoring insights.
A government agency's public records search system processing 50,000 new documents monthly implemented automated quality monitoring: (1) Random sampling of 1% of daily chunked output for automated analysis; (2) Coherence scoring using a fine-tuned BERT model that predicts whether chunks contain complete semantic units; (3) Boundary quality checks that flag chunks ending mid-sentence or splitting obvious semantic units (detected through pronoun reference analysis); (4) Weekly retrieval testing using a static test set of 200 queries with known relevant documents; (5) Anomaly detection that identifies document types with significantly lower quality scores. This monitoring revealed that scanned documents from the 1970s-1980s had 40% lower coherence scores due to OCR errors, prompting the team to implement enhanced OCR preprocessing and specialized chunking logic for historical documents. The monitoring system enabled quality maintenance at scale without manual review of millions of chunks.
Challenge: Balancing Storage Costs with Retrieval Quality
Smaller chunks generally improve retrieval precision but increase vector database storage requirements and potentially query latency, while larger chunks reduce costs but may degrade retrieval quality 13. Organizations must balance these competing concerns within budget and performance constraints, particularly when processing large document collections 8.
Solution:
Conduct cost-benefit analysis that quantifies the relationship between chunk size, storage costs, query performance, and retrieval quality for the specific use case, then implement adaptive strategies that optimize this trade-off 25. Consider hybrid approaches that use different chunk sizes for different document types based on their retrieval importance and query frequency.
A healthcare organization's clinical knowledge base faced storage costs of $12,000 monthly for 60 million vectors generated from 400-token chunks of medical literature and clinical guidelines. The team conducted analysis revealing that: (1) Clinical guidelines (10% of documents) received 70% of queries and required high precision; (2) Background medical literature (90% of documents) received 30% of queries and tolerated moderate precision. They implemented a tiered approach: clinical guidelines maintained 400-token chunks for precision, while background literature moved to 700-token chunks. This reduced vector count to 38 million, cutting storage costs to $7,600 monthly (37% reduction) while retrieval quality testing showed only 3% precision decrease for background literature queries and no degradation for high-priority clinical guideline queries. The cost-benefit analysis demonstrated that the modest precision trade-off for low-frequency queries justified the substantial cost savings.
References
- arXiv. (2023). Dense Passage Retrieval for Open-Domain Question Answering. https://arxiv.org/abs/2307.03172
- arXiv. (2020). Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks. https://arxiv.org/abs/2005.11401
- ACL Anthology. (2023). Semantic Chunking for Long Document Retrieval. https://aclanthology.org/2023.acl-long.868/
- Google Research. (2019). BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. https://research.google/pubs/pub46826/
- arXiv. (2021). Hierarchical Document Representations for Efficient Information Retrieval. https://arxiv.org/abs/2104.08663
- ACL Anthology. (2022). Improving Dense Retrieval through Text Segmentation Strategies. https://aclanthology.org/2022.naacl-main.291/
- arXiv. (2022). Multi-Resolution Document Chunking for Neural Information Retrieval. https://arxiv.org/abs/2212.10496
- ScienceDirect. (2023). Document Segmentation Strategies in Modern Information Retrieval Systems. https://www.sciencedirect.com/science/article/pii/S0306457323001115
