Make high-stakes decisions with receipts. Mavera turns live signals into market proof — a Listen / Decide / Make / Measure platform built for leaders who need more than generic AI. A synthetic audience engine blends generative models with live context, simulating audience reactions before budget is committed and producing explainable, defensible insight (research benchmarked at 98% agreement with Harvard's OASIS human-subject study). Under the hood: 95+ database models, 100+ API routes, 40+ Plate editor plugins, and 15+ third-party integrations spanning conversational AI, document generation, meeting intelligence, focus groups, video analysis, and news digestion.
Mavera (mavera.io) is a production AI research and insights platform organized around a four-pillar framework — Listen (stream signals that impact success), Decide (simulate audience reactions before committing budget), Make (create work that connects now and is defensible later), and Measure (study impact while there's still time to adjust). The synthetic audience engine blends generative models with live context to handle complex, sensitive conversations where accuracy and accountability matter, with transparent reasoning and traceable evidence on every output (98% agreement with Harvard's OASIS human-subject study). The platform consolidates seven heavyweight workflows into a single workspace: persona-based conversational AI, AI-streamed document generation through editable templates, meeting intelligence powered by Recall.ai with AI-extracted highlights and tasks, structured focus groups with 11 question types, frame-by-frame video analysis, scheduled news digests with persona-specific perspectives, and the Mave agent for branched, evidence-tracked research. The system spans 95+ Prisma models, 100+ API routes, 61 server-action files, 40+ Plate editor plugins, and 15+ third-party integrations — delivered as a multi-tenant Next.js 14 App Router application running on PostgreSQL, AWS, and Trigger.dev with full Auth0 RBAC across 11 workspace roles.
A single Plate v44 editor instance powers two distinct surfaces: inline artifact editing in chat (artifacts parsed from <artifact> XML tags in streamed responses) and the full /generate template editor. 40+ plugins cover headings, tables, multi-column layouts, callouts, code with Prism highlighting, mentions, comments, equations, Excalidraw diagrams, embedded media, and AI/copilot commands. Auto-save runs through PlateEditorWrapper with 1s debounce, and serialization plugins handle DOCX import, Markdown round-tripping, HTML for email, and CSS inlining via Juice.
Every chat response is a Zod-validated object: response text, optional editable artifact, emotion (valence/arousal), pragmatic (realism/actionability), biases array with severity, contextual factors, future considerations, confidence level, opinion spread, perspective shifts, and follow-up questions. The schema is enforced via generateObject() so every persona response stays analyzable and renderable in the UI without client-side parsing fallbacks.
Recall.ai bots join Google Meet / Zoom calls, record audio + video, and emit transcripts. A Trigger.dev job (analyze-meeting, 25KB, concurrency 5) parses VTT/SRT, then runs structured AI extraction for: 7 highlight categories (KEY_INSIGHT, OBJECTION, COMMITMENT, QUESTION, RISK, OPPORTUNITY, ACTION_ITEM), priority-tagged tasks with owners and due dates, decisions with deadlines, transcript-span evidence linked back to every artifact, coaching metrics (talk ratio, question rate, interruptions, sales quality), and optional custom-schema extraction with 9 field types.
Mave is a stateful research agent with conversation branching, evidence tracking, and tool execution. The orchestrator (mave-orchestrator.ts, 60KB) coordinates LLM calls, tool execution (Tavily search/extract, CircleMind KB queries, SEMrush, image/video generation), and branch management (mave-branches.ts, 20KB). Output is presented across four tabs — Answer, Evidence (citation-linked claims), Timeline (tool-call history), and Flagged — with per-thread persona attachment.
Account → Workspace → Project → Thread → Message hierarchy enforced via Prisma relations and middleware role checks. Each workspace owns its own brand voice, persona pool, integrations, dashboard layout (react-grid-layout JSON), budget alerts, knowledge base, and credit usage tracking (WorkspaceUsage daily aggregation, ProjectUsage per-project). Workspace types span USER, STARTER, PROFESSIONAL_AGENCY, ENTERPRISE, MARKETING_AGENCY, and BASIC_AGENCY.
9 background jobs run on Trigger.dev with prismaExtension and ffmpegExtension: meeting analysis (queue concurrency 5), video chunked frame analysis, focus-group response generation (concurrency 10), brand-voice extraction (concurrency 3), knowledge-base building via Perplexity + CircleMind, news-digest cron (every 30 min), multi-step onboarding workflow, and daily workspace analytics cron at 2 AM UTC (32KB job). Retries are exponential up to 3 attempts.
middleware.ts runs on every protected request: Auth0 session check, public-route allowlist, invite-route bypass, subscription gate redirecting to /subscription, and dev-mode role injection. RBAC is matrix-based across 11 roles × task groups (FOCUS_GROUP, VIDEO_ANALYSIS, WORKSPACE, PROJECTS, CHATS, PERSONAS, ASSETS, WORKSPACE_SETTINGS, API_KEYS) and checked via isAllowed(role, permission, taskName) at every server action and API route boundary.
5-format export pipeline: DOCX via remark-docx, PDF via html2canvas + pdf-lib, HTML via Plate's HTML serializer with embedded styles, PNG via html2canvas, and Markdown via Plate's Markdown plugin — exposed from the editor toolbar, chat artifacts, generation outputs, meeting analyses, and focus-group results. The integrations layer (80KB) handles OAuth flows and data push to Slack (Block Kit), Linear, HubSpot, Salesforce, Asana, Monday.com, and Jira, plus Google Drive / OneDrive / Dropbox file pickers.
Stream chat with pre-built or custom personas, including MCDA-weighted custom personas with localization and purpose packs
Database-driven template grid with dynamic input fields, brand-voice injection, and AI streaming directly into the Plate editor
Recall.ai bots that join meetings, transcribe, then extract highlights, tasks, decisions, coaching metrics, and evidence
Stripe-billed AI research sessions with 11 question types and aggregated insights across configurable sample sizes
S3-backed chunked frame extraction with multi-modal AI scoring across emotional, cognitive, behavioral, speech, and visual axes
Scheduled persona-aware news digests via Perigon, delivered through Resend with React Email templates
Branched research agent with tool execution, evidence tracking, and Answer/Evidence/Timeline/Flagged views
Custom meeting-extraction schemas with 9 field types, publishable for community sharing and reuse
One-click push of meeting outputs and tasks to Slack, Linear, HubSpot, Salesforce, Asana, Monday.com, and Jira







