Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.moorcheh.ai/llms.txt

Use this file to discover all available pages before exploring further.

Moorcheh Enterprise Changelog

Track enterprise-specific releases in one place, including access control, metadata filtering, enterprise retrieval controls, and ingestion architecture updates.
Version 1.1.0
Released April 20, 2026

Platform parity

  • Changes in this release are applied consistently on GCP and AWS enterprise deployments.

Backend observability and Moorcheh dashboard

We implemented backend observability (logging and monitoring in the services) and surfaced it in the Moorcheh dashboard so enterprise admins (and operators with dashboard access) can review the same signals without jumping only to cloud consoles.
  • Dashboard views - Log and monitoring views in the Moorcheh dashboard show the latest information from the backend when you open or refresh the page. Up to 30 days of log and monitoring history is available for review (enterprise lookback / retention).

Logging

  • Structured logging - Every significant backend action is logged for request handling, async jobs, and integration points.
  • Dashboard visibility - Admins can inspect and follow those logs from the Moorcheh dashboard for troubleshooting, audits, and day-to-day operational awareness.

Monitoring and alarms

  • Runtime metrics - Per-function runtime with percentile breakdowns (p50, p95, p99) where applicable, viewable in the dashboard alongside other health signals.
  • Database health - Alarms cover database latency, errors, and capacity signals; admins can see monitoring state and drill-downs in the Moorcheh dashboard.
  • Broad coverage - More than 60 alarms across the platform to catch regressions, saturation, and abnormal traffic patterns early.

Per-service summaries (in dashboard)

In the Moorcheh dashboard, summaries are grouped by major service so admins can focus on the component that matters:
  • AWS - Amazon DynamoDB, SQS, API Gateway, Lambda (and related integration points).
  • GCP - Firestore, Pub/Sub, API Gateway, Cloud Run (and related integration points).
Version 1.0.0
Released April 1, 2026

Deployment footprint

  • Current enterprise deployments run on both AWS and GCP.

User management and access control

  • Admin can create namespaces.
  • Admin can delete namespaces, including async delete flows for large namespaces.
  • Admin can invite users and list invited users.
  • Admin can grant namespace access to specific users.
  • Admin can revoke namespace access.
  • Admin can remove users when needed.

Namespace access model

  • Namespaces can be owned by a user/admin and shared to other users.
  • Access is managed at namespace granularity.
  • Search and answer flows honor namespace-level access boundaries.
  • This enables secure multi-tenant collaboration and controlled sharing.

Namespace lifecycle and data management

  • Create namespace.
  • List namespaces.
  • Delete namespace.
  • Delete namespace data.
  • Get namespace data/documents.
  • Document status tracking.

Metadata capabilities

File-level metadata management

  • Metadata is supported in file upload, upload-url, text upload, and vector upload flows.
  • Metadata can be updated for existing documents/items after ingestion.
  • Supported metadata types include string, number, boolean, and array.
  • Metadata values support spaces and most special characters.
  • # is excluded for indexed filtering use-cases.

Metadata filter feature (search-time filtering)

  • Search supports structured metadata filters.
  • Per-filter matching mode (match) controls multi-value evaluation inside a filter.
  • Cross-filter combination mode (combine) controls evaluation across filters.
  • Logical patterns support both any and all.

Metadata as a sharing primitive

  • Metadata-driven filtering can model sharing and visibility logic (team scope, ownership hints, publish flags, access tags).
  • This enables consistent retrieval enforcement in search and answer experiences.

Search and RAG enterprise controls

  • In addition to core platform search, enterprise supports metadata-constrained retrieval and policy-like filtering behavior.
  • Inline keyword constraints with #keyword are supported as a refinement stage.
  • One-call answer APIs support retrieval + context assembly + generation with namespace-scoped grounding.
  • Model selection controls can be applied by deployment and governance needs.

Region and model flexibility

  • Current AWS embedding model is Cohere Embed v4 (cohere.embed-v4, runtime variant cohere.embed-v4:0 on Bedrock).
  • AI models can be changed based on enterprise requirements.
  • AWS deployments can choose Bedrock AI models by target region.
  • GCP deployments can also select models based on regional and customer requirements.
  • This supports compliance, latency, and cost/performance alignment.

Large file upload (up to 5 GB) via upload URL

  1. Client requests upload URL from API (upload-url endpoint).
  2. API returns pre-signed S3 PUT URL and storage target details.
  3. Client uploads file directly to object storage.
  4. Backend processing pipeline ingests file asynchronously.
  5. Chunks, embeddings, and metadata are written to data stores for search and RAG.

Processing pipeline overview

  • Document, text, and vector ingestion jobs run asynchronously.
  • Embedding and vector preparation are handled in worker flows.
  • Chunk persistence is handled in worker flows.
  • Namespace/data deletion jobs and document status updates run asynchronously.

Implementation coverage

  • Enterprise features are implemented across both Node.js and Rust service stacks.
  • Coverage includes user/access management, namespace lifecycle, ingestion, metadata update/filter flows, enterprise retrieval controls, answer workflows, and upload-url large-file support.

Feature summary for enterprise stakeholders

  • Admin-led user and namespace governance.
  • Namespace sharing and access control.
  • Full metadata lifecycle support (attach, update, filter).
  • Advanced metadata filtering logic for retrieval control.
  • Enterprise retrieval controls layered on top of platform search capabilities.
  • One-call RAG with configurable AI model selection by region/choice.
  • Large-file direct upload flow (up to 5 GB).
  • Parallel Node.js and Rust implementation coverage for core platform capabilities.