# AI Models

# AI Models

Curiosity Workspace uses AI models in three common ways:

  • Understanding: NLP pipelines that extract structure from text (entities, signals, links).
  • Retrieval: embeddings used for semantic similarity (vector search) and re-ranking.
  • Generation: LLMs used for synthesis, assistance, and workflow automation.

The important architectural point: AI features are most reliable when they are grounded in your workspace data via graph + search retrieval.

# Embedding models (semantic similarity)

Embedding models map text (and sometimes other modalities) into vectors. In Curiosity Workspace, embeddings are used for:

  • vector search (find semantically similar items)
  • clustering and similarity (group related items)
  • candidate generation for AI-assisted workflows

Design considerations:

  • choose which fields get embeddings (usually long, descriptive text)
  • enable chunking when fields exceed model context limits
  • decide whether vector search is a primary retrieval method or a supplement to text search

See NLP → Embeddings and Search → Vector Search.

# NLP pipelines (extraction + enrichment)

NLP pipelines transform raw text into structured outputs, such as:

  • extracted entities (people, products, IDs)
  • normalized tokens and language-specific parsing
  • optional entity linking into your graph

This enables:

  • better filters (entities become facets)
  • better graph navigation (mentions → resolved entities)
  • better retrieval grounding for LLMs

See NLP → Overview and NLP → Entity Extraction.

# LLMs (generation + orchestration)

LLMs are typically used to:

  • answer questions using retrieved context
  • summarize, classify, or extract structured outputs
  • drive multi-step workflows (tools, endpoints, actions)

Recommended patterns:

  • retrieval first: fetch relevant nodes/documents before prompting
  • tooling: move business logic into endpoints/tasks rather than relying on prompts alone
  • auditability: store inputs/outputs where needed (policy dependent)

See AI & LLMs → Overview and AI & LLMs → Prompting Patterns.

# Safety and governance (conceptual)

Production AI typically needs:

  • permission-aware retrieval
  • logging/auditing for sensitive workflows
  • strict separation of admin-only capabilities

See Administration → Security.

# Next steps