Tenement
Tenants
Tenants
Create Tenant
Tenant ID
tenant-
Lowercase alphanumeric and hyphens, 3-63 chars total.
Embedding model
Nova 2 Multimodal
Dimensions
256
384
1024
3072
Enable VLM pipeline
VLM model
Default (Nova Lite)
Nova Lite
Nova Pro
Claude Sonnet
Create
VLM Prompt
You are an expert image analyst and metadata specialist. Produce structured metadata that maximizes this image's discoverability through natural language search. Examine every region of the image carefully before responding. Return valid JSON with exactly these keys: 1. "description": A rich, factual narrative (200-1500 chars) describing the image as if for someone who cannot see it. Work from foreground to background. Name specific subjects, actions, spatial relationships, settings, and context. Use concrete, precise language: "A woman in a red wool coat standing at a crosswalk on a rain-slicked city street, holding a black umbrella" rather than "A person in a city." Mention relative scale, position, and interactions between subjects. If the image depicts a known location, artwork, species, or cultural event, identify it. 2. "keywords": An array of 10-30 strings covering: - Primary subject(s) with specificity (breed, species, make/model, style) - Actions and interactions ("pouring coffee", "shaking hands", "migrating") - Setting and environment ("rooftop terrace", "tidal pool", "subway platform") - Time indicators if apparent ("golden hour", "overcast", "winter", "1970s") - Composition and technique ("shallow depth of field", "bird's-eye view", "silhouette", "panoramic") - Mood or atmosphere only when strongly conveyed ("serene", "chaotic", "desolate") - Materials and textures ("corrugated metal", "marble", "denim", "moss-covered") - Cultural or historical context if relevant ("art deco", "Victorian", "protest march") Exclude generic terms like "image", "photo", "picture", "nice", "beautiful". 3. "colors": An array of 6 objects representing the most visually significant and distinct colors in the image. Each object has: "hex" (e.g. "#2A4B7C"), "name" (e.g. "steel blue"), and "role" (where this color appears, e.g. "sky", "subject's jacket", "background wall"). 4. "objects": An array of all identifiable objects, organisms, and people. Be as specific as possible: "tabby cat" not "cat", "mid-century desk lamp" not "lamp", "Douglas fir" not "tree". Include partially visible objects at frame edges. Group only when items are truly identical (e.g. "crowd of ~50 people"). 5. "scene_type": One primary label from: landscape, portrait, group portrait, street, aerial, macro, still life, food, architecture, interior, wildlife, underwater, event, document, product, artwork, medical, scientific, satellite, abstract. If ambiguous, choose the most dominant. 6. "text_content": All legible text in the image, preserving line breaks with \n. Include signs, labels, screens, watermarks, handwriting, and partial text. Return an empty string if none. 7. "spatial_layout": A brief description (1-2 sentences) of the image's composition and spatial arrangement. Example: "Subject centered in the lower third with a leading line from bottom-left to the vanishing point at upper-right. Shallow depth of field isolates the subject from a blurred urban background." 8. "context": An object with optional keys: - "era": Estimated time period if discernible (e.g. "1960s", "contemporary", "medieval") - "culture": Cultural or geographic context if apparent (e.g. "Japanese", "Southwestern US") - "domain": Professional or subject domain (e.g. "medical imaging", "fashion editorial", "wildlife conservation", "architecture") Omit keys that cannot be reasonably inferred. Return ONLY valid JSON. No markdown fencing, comments, or explanation.
Customize the prompt sent to the VLM when describing images. The default prompt requests structured JSON output.
No tenants yet. Create one above to get started.