#
Embeddings
#
Embeddings
Embeddings are vector representations of content that enable semantic similarity and vector search. In Curiosity Workspace, embeddings are commonly used to:
- power vector search (semantic retrieval)
- find similar items (related cases/documents)
- provide candidate context for LLM workflows
#
How embeddings are used
Typical workflow:
- Choose which fields should be embedded (usually longer, descriptive text).
- Configure embedding index creation for those fields.
- Run similarity queries to retrieve nearest neighbors for a user query or a node’s content.
#
Choosing what to embed
Good candidates:
- description/body fields
- conversations and transcripts
- summaries (if they contain meaningful signal)
Poor candidates:
- IDs and codes
- short labels (often better handled by keyword search)
#
Chunking (important for long text)
If a field can be very long:
- enable chunking
- ensure chunks align with semantic units (paragraphs, messages)
Chunking typically improves recall but may require careful tuning so results remain interpretable.
#
AI Configuration Guide
#
Selecting Embedding Models
- Dimensionality: Models with higher dimensions (e.g., 1536) generally capture more semantic detail but require more memory and storage.
- Language Support: Choose multilingual models if your data spans multiple languages.
- Provider: Decide between local models (running on your infrastructure) and cloud-hosted models (via OpenAI, Azure, or Anthropic).
#
Memory and Resource Consumption
- Index Size: Vector indexes can consume significant memory. Ensure your server has enough RAM to keep active indexes in memory for fast retrieval.
- Storage: SSD storage is highly recommended to minimize latency during index updates and queries.
#
Integration with LLMs
Embeddings are the foundation for Retrieval-Augmented Generation (RAG). By retrieving the most semantically relevant chunks from the workspace, you provide the LLM with the high-quality context it needs to generate accurate and grounded responses.
#
Evaluation
To validate embeddings:
- prepare a list of “similarity questions” (e.g., “find similar incidents”)
- confirm results are relevant and diverse
- tune cutoffs (how similar is “similar enough”?)
#
Next steps
- Use embeddings for retrieval: Search → Vector Search
- Combine with keyword retrieval: Search → Hybrid Search
See also
Curiosity Workspace uses AI models in three common ways:
This page documents the conceptual API surface for embeddings and semantic similarity in Curiosity Workspace.
Entity extraction identifies meaningful spans of text (entities) and turns them into structured outputs. Examples:
Tuning your Natural Language Processing (NLP) pipelines is essential for high-quality entity extraction and search relevance.
Natural Language Processing (NLP) in Curiosity Workspace turns raw text into structured signals you can search, filter, and connect to your graph.
This quickstart takes you through the smallest “end-to-end loop” that demonstrates the Curiosity Workspace value chain:
Curiosity Workspace search is designed for structured + unstructured data, and integrates tightly with the graph model.
Semantic similarity answers “how related are these two items by meaning?” even when they do not share keywords.
This page collects AI patterns shown in the demo repository, rewritten as reusable building blocks for Curiosity Workspace applications.