The Prompt Engineering Project

Documentation

Everything you need to get started with prompt libraries.

Quick Start

Three steps to your first prompt library execution.

Install the IO prompt library SDK and dependencies:

# Install the core package
npm install @io/prompt-library

# Install Notion integration
npm install @io/notion-connector

# Install the CLI
npm install -g @io/cli

Getting Started Guide

Follow these five steps to build your first prompt library from scratch.

1. Define Your Context Brief

Start by creating a Context Brief — a single-source-of-truth document that captures everything about your company: mission, values, target audience, tone of voice, and competitive positioning. This document powers every AI interaction and ensures consistent, on-brand output across all prompts.

2. Complete the Questionnaire

Walk through the guided questionnaire that extracts the critical information needed for your Context Brief. Each question is designed to surface insights that most brands overlook — from nuanced positioning statements to audience psychographics and brand personality traits.

3. Connect Your Notion Database

Link your Notion workspace and select the database where your Prompt Library lives. The connector automatically maps column prompts to database properties and configures AI Auto-Fill triggers for each field.

4. Configure Column Prompts

Each column in your Notion database gets its own structured prompt that references your Context Brief. With 23 column prompts for Company Identity alone, every piece of generated content is deeply contextualized and aligned with your brand strategy.

5. Execute Fan-Out, Fan-In

Trigger all column prompts simultaneously with a single action. The fan-out architecture fires 23 parallel prompts from one Context Brief, then fans results back into your structured database — transforming hours of manual work into seconds of automated generation.

API Reference

Four core APIs power the entire prompt engineering platform.

Context Brief API

The Context Brief API lets you programmatically create, read, update, and manage context briefs. Load briefs from Markdown files, sync with Notion pages, and validate completeness scores. Supports versioning so you can track how your brand context evolves over time.

Prompt Library API

Access and execute any prompt from your library via the API. Supports single execution, batch processing, and the full fan-out/fan-in parallel architecture. Each prompt automatically injects your Context Brief for maximum relevance and brand alignment.

Knowledge Base API

Query and update your Knowledge Base programmatically. The API supports semantic search across all entries, automatic categorization, and real-time sync with your Notion databases. Build living knowledge systems that grow smarter with every interaction.

Auto-Fill API

The Auto-Fill API powers Notion's Custom AI Auto-Fill integration. Configure triggers, map column prompts to database properties, and control execution order. Supports conditional logic so prompts can reference outputs from other columns in the same row.

Architecture Overview

Four layers power the prompt engineering platform.

Changelog from my journey

I've been working on Aceternity for the past 2 years. Here's a timeline of my journey.

Layer 1: Context

The foundation layer. Context Briefs serve as the single source of truth — capturing company identity, audience insights, brand voice guidelines, and competitive positioning in a structured document that every downstream prompt references.

Context Layer

Layer 2: Prompts

The execution layer. Structured column prompts live in Notion databases and are designed to generate specific outputs. Each prompt references the Context Brief and follows a consistent template: role, context injection, specific instruction, output format, and quality criteria.

Prompts Layer

Layer 3: Orchestration

The parallelization layer. Fan-out architecture fires all 23 column prompts simultaneously from a single Context Brief. Results fan back into the structured database, with dependency resolution ensuring prompts that reference other outputs execute in the correct order.

Orchestration Layer

Layer 4: Knowledge

The intelligence layer. Generated outputs feed back into the Knowledge Base, creating a living system that grows smarter with every execution. Semantic indexing enables cross-referencing and retrieval, turning isolated outputs into interconnected business intelligence.

Knowledge Layer

Component Documentation

Six core components make up the IO prompt engineering system.

Context Brief

The foundational document that captures your company's identity, audience, voice, and strategy. Every prompt references this single source of truth to ensure consistent, on-brand AI outputs.

Questionnaire

A guided intake form that extracts critical business information through targeted questions. Responses auto-populate your Context Brief with structured, prompt-ready data.

Prompt Library

A Notion database of structured prompts organized by category. Each prompt follows a consistent template with role, context injection, instruction, output format, and quality criteria.

Knowledge Base

A living repository of generated content and curated insights. Supports semantic search, automatic categorization, and real-time sync with your Notion workspace.

Column Prompts

Individual prompts mapped to specific Notion database columns. Designed for AI Auto-Fill, each column prompt generates targeted content for a single property in your database.

Chat Interface

Conversational access to your entire prompt library and knowledge base. Ask questions, generate content, or explore your brand context through natural language interaction.

Start Building

Create your first prompt library today. From context brief to full execution in minutes.

IOIOIOtm

The Prompt Engineering Project by Intelligent Operations