Create Index
Trigger indexing for a parsed document to make it searchable.Endpoint
POST /api/index
Request Body
The ID of the document to index (must be already parsed)
The embedding model to use for vectorization
Size of text chunks for embedding (in tokens)
Overlap between consecutive chunks (in tokens)
Response
Unique identifier for the indexing job
Document being indexed
Indexing status:
pending, processing, completed, failedNumber of chunks created from the document
Example
Get Index Status
Check the status of an indexing job and retrieve statistics.Endpoint
GET /api/index/{index_id}
Path Parameters
The indexing job identifier
Response
Indexing job identifier
Document being indexed
Current status:
pending, processing, completed, failedTotal number of chunks created
Number of vectors successfully embedded
Completion percentage (0-100)
Error message if indexing failed
Example
Re-index Document
Re-index an existing document with new settings or after content updates.Endpoint
PUT /api/index/{document_id}
Request Body
Force re-indexing even if document hasn’t changed
Update the embedding model
Example
Embedding Models
OpenAI
text-embedding-3-large(3072 dimensions)text-embedding-3-small(1536 dimensions)
Custom
- Bring your own embedding model
- Contact support for integration
Chunking Strategies
Fixed-size chunking
Fixed-size chunking
Default strategy that splits text into fixed-size chunks with overlap. Best for general-purpose indexing.Recommended settings:
- Chunk size: 512 tokens
- Overlap: 50 tokens
Semantic chunking
Semantic chunking
Splits text based on semantic boundaries (paragraphs, sections). Better for preserving context.
Available in Pro and Enterprise plans.
