mirror of
https://github.com/garrytan/gstack.git
synced 2026-05-08 21:49:45 +08:00
* feat: gstack-gbrain-mcp-verify helper for remote MCP probe
Probes a remote gbrain MCP endpoint with bearer auth. POSTs initialize,
classifies failures into NETWORK / AUTH / MALFORMED with one-line
remediation hints, and runs a tools/list capability probe to detect
sources_add MCP support (forward-compat for when gbrain ships URL ingest).
Token consumed from GBRAIN_MCP_TOKEN env, never argv. Required to set
both 'application/json' AND 'text/event-stream' in Accept; that gotcha
costs 10 minutes of debugging when missed (regression-tested).
Live-verified against wintermute (gbrain v0.27.1).
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
* feat: gstack-artifacts-init + gstack-artifacts-url helpers
artifacts-init replaces brain-init with provider choice (gh / glab /
manual), per-user gstack-artifacts-$USER repo, HTTPS-canonical storage in
~/.gstack-artifacts-remote.txt, and a "send this to your brain admin"
hookup printout. Always prints the command, never auto-executes — gbrain
v0.26.x has no admin-scope MCP probe (codex Finding #3).
artifacts-url centralizes HTTPS↔SSH/host/owner-repo conversion so callers
don't each string-mangle (codex Finding #10). The remote-conflict check in
artifacts-init compares at the canonical level so re-running with HTTPS
input doesn't trip on a stored SSH URL for the same logical repo.
The "URL form not supported" branch prints a two-line clone-then-path
form for gbrain v0.26.x; the supported branch is a one-liner with --url
ready for when gbrain ships URL ingest.
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
* feat: extend gstack-gbrain-detect with mcp_mode + artifacts_remote
Adds two new fields to detect's JSON output:
- gbrain_mcp_mode: local-stdio | remote-http | none
Resolved via 3-tier fallback (codex Finding D3): claude mcp get --json
→ claude mcp list text-grep → ~/.claude.json jq read. If Anthropic moves
the file format, the first two tiers absorb it.
- gstack_artifacts_remote: HTTPS URL from ~/.gstack-artifacts-remote.txt
Falls back to ~/.gstack-brain-remote.txt during the v1.27.0.0 migration
window so detect doesn't return empty between upgrade and migration.
Existing detect tests still pass (15/15). New 19 tests cover every fallback
tier independently, plus a schema regression for /sync-gbrain compat.
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
* feat: setup-gbrain Path 4 (remote MCP) + artifacts rename
Path 4 lets users paste an HTTPS MCP URL + bearer token and registers it
as an HTTP-transport MCP without needing a local gbrain CLI install. The
flow:
- Step 2 gains a fourth option (Remote gbrain MCP)
- Step 4 adds Path 4 sub-flow: collect URL, secret-read bearer, verify
via gstack-gbrain-mcp-verify (NETWORK / AUTH / MALFORMED classifier)
- Step 5 (local doctor), Step 7.5 (transcript ingest), Step 5a's stdio
branch all skip on Path 4
- Step 5a adds an HTTP+bearer registration form: claude mcp add
--transport http --header "Authorization: Bearer ..."
- Step 7 renamed "session memory sync" → "artifacts sync" and now calls
gstack-artifacts-init (which always prints the brain-admin hookup
command — no auto-execute, codex Finding #3)
- Step 8 CLAUDE.md block branches: remote-http includes URL + server
version (never the token); local-stdio keeps engine + config-file
- Step 9 smoke test on Path 4 prints the curl-equivalent for
post-restart verification (MCP tools aren't visible mid-session)
- Step 10 verdict block has separate templates per mode
Idempotency: re-running with gbrain_mcp_mode=remote-http already in
detect output skips Step 2 entirely and goes to verification.
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
* refactor: rename gbrain_sync_mode → artifacts_sync_mode (v1.27.0.0 prep)
Hard rename, no dual-read alias (codex Finding D4). The on-disk migration
script (Phase C, separate commit) renames the config key in users'
~/.gstack/config.yaml and any CLAUDE.md blocks.
Touched call sites:
- bin/gstack-config defaults + validation + list/defaults output
- bin/gstack-gbrain-detect (gstack_brain_sync_mode field still emitted
with the same name for downstream-tool compat; reads new key)
- bin/gstack-brain-sync, bin/gstack-brain-enqueue, bin/gstack-brain-uninstall
- bin/gstack-timeline-log (comment ref)
- scripts/resolvers/preamble/generate-brain-sync-block.ts: renames key,
branches on gbrain_mcp_mode=remote-http to emit "ARTIFACTS_SYNC:
remote-mode (managed by brain server <host>)" instead of the local
mode/queue/last_push line (codex Finding #11)
- bin/gstack-brain-restore + bin/gstack-gbrain-source-wireup: read
~/.gstack-artifacts-remote.txt with ~/.gstack-brain-remote.txt fallback
during the migration window
- bin/gstack-artifacts-init: tolerant of unrecognized URL forms (local
paths, file://, self-hosted gitea) so test infrastructure and unusual
remotes work without canonicalization
- test/brain-sync.test.ts: gstack-brain-init → gstack-artifacts-init
- test/skill-e2e-brain-privacy-gate.test.ts: artifacts_sync_mode keys
- test/gen-skill-docs.test.ts: budget 35K → 36.5K for the new MCP-mode
probe in the preamble resolver
- health/SKILL.md.tmpl, sync-gbrain/SKILL.md.tmpl: comment + verdict line
Hard delete:
- bin/gstack-brain-init (replaced by bin/gstack-artifacts-init in v1.27.0.0)
- test/gstack-brain-init-gh-mock.test.ts (replaced by gstack-artifacts-init.test.ts)
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
* chore: regenerate SKILL.md files after artifacts-sync rename
Mechanical regen via \`bun run gen:skill-docs --host all\`. All */SKILL.md
files reflect the renamed config key (gbrain_sync_mode →
artifacts_sync_mode), the renamed remote-helper file
(~/.gstack-artifacts-remote.txt with brain fallback), the renamed init
script (gstack-artifacts-init), and the new ARTIFACTS_SYNC: remote-mode
status line that fires when a remote-http MCP is registered.
Golden fixtures (test/fixtures/golden/*-ship-SKILL.md) refreshed to match
the regenerated default-ship output.
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
* feat: v1.27.0.0 migration — gstack-brain → gstack-artifacts rename
Journaled, interruption-safe migration. Six steps, each writes to
~/.gstack/.migrations/v1.27.0.0.journal on success; re-entry resumes
from the next un-done step. On final success, journal is replaced by
~/.gstack/.migrations/v1.27.0.0.done.
Steps:
1. gh_repo_renamed gh/glab repo rename gstack-brain-$USER →
gstack-artifacts-$USER (idempotent: detects
already-renamed and skips)
2. remote_txt_renamed mv ~/.gstack-brain-remote.txt → artifacts file,
rewriting URL path to match the new repo name
3. config_key_renamed sed -i in ~/.gstack/config.yaml flips
gbrain_sync_mode → artifacts_sync_mode
4. claude_md_block sed flips "- Memory sync:" → "- Artifacts sync:"
in cwd CLAUDE.md and ~/.gstack/CLAUDE.md
5. sources_swapped gbrain sources add NEW (verify) → remove OLD
(codex Finding #6: add-before-remove ordering,
no downtime window). On remote-MCP mode, prints
commands for the brain admin instead of executing.
6. done touchfile + delete journal
User opt-out: any "n" or "skip-for-now" answer at the initial prompt
writes a marker file that prevents re-prompting; user can re-invoke
via /setup-gbrain --rerun-migration.
11 unit tests cover: nothing-to-migrate, GitHub happy path, idempotent
re-run, journal-resume mid-flight, remote-MCP print-only path,
add-before-remove ordering verification, add-fail → old source stays
registered, CLAUDE.md field rewrite.
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
* test: regression suite + E2E for v1.27.0.0 rename
Three new regression tests guard the rename's blast radius (per codex
Findings #1, #8, #9, #12):
- test/no-stale-gstack-brain-refs.test.ts: greps bin/, scripts/, *.tmpl,
test/ for forbidden identifiers (gstack-brain-init, gbrain_sync_mode);
fails CI if any non-allowlisted file references them.
- test/post-rename-doc-regen.test.ts: confirms gen-skill-docs output has
no stale references in any */SKILL.md (the cross-product blind spot).
- test/setup-gbrain-path4-structure.test.ts: structural lint over the
Path 4 prose contract — STOP gates after verify failure, never-write-
token rules, mode-aware CLAUDE.md block, bearer always via env-var.
Two new gate-tier E2E tests (deterministic stub HTTP server, fixed inputs):
- test/skill-e2e-setup-gbrain-remote.test.ts: Path 4 happy path. Stubs
an HTTP MCP server, drives the skill via Agent SDK with a stubbed
bearer, asserts claude.json gets the http MCP entry, CLAUDE.md gets
the remote-http block, the secret token NEVER leaks to CLAUDE.md.
- test/skill-e2e-setup-gbrain-bad-token.test.ts: stub server returns 401;
asserts the AUTH classifier hint surfaces, no MCP registration occurs,
CLAUDE.md is unchanged. Regression guard for the "verify failed → STOP"
rule.
touchfiles.ts: setup-gbrain-remote and setup-gbrain-bad-token added at
gate-tier so CI catches Path 4 regressions on every PR.
Plus a few comment refs flipped: bin/gstack-jsonl-merge, bin/gstack-timeline-log
(legacy gstack-brain-init mentions in headers).
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
* release: v1.27.0.0 — /setup-gbrain Path 4 + brain → artifacts rename
Bumps VERSION 1.26.4.0 → 1.27.0.0 (MINOR per CLAUDE.md scale-aware bump
guidance: ~1500 line net change including a new path in /setup-gbrain,
two new bin helpers, a journaled migration, 59 new tests, and a config
key rename across the codebase).
CHANGELOG entry covers: Path 4 (Remote MCP) end-to-end, the brain →
artifacts rename, the journaled migration, the verify-helper error
classifier, the artifacts-init multi-host provider choice. Includes
the canonical Garry-voice headline + numbers table + audience close
per the release-summary format.
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
* test: demote setup-gbrain Path 4 E2E to periodic-tier
The Agent SDK E2E tests for Path 4 (skill-e2e-setup-gbrain-remote and
skill-e2e-setup-gbrain-bad-token) are inherently non-deterministic —
the model interprets "follow Path 4 only" prompts flexibly and can
skip Step 8 (CLAUDE.md write) or shortcut past the verify helper, which
makes the gate-tier assertions flaky.
The deterministic gate coverage for Path 4 is in
test/setup-gbrain-path4-structure.test.ts: a fast structural lint that
catches AUQ-pacing regressions and prose contract drift in <200ms with
zero token spend. That test is the right tool for catching the failure
mode the gate-tier was meant to guard against.
The Agent SDK E2E tests stay available on-demand for periodic-tier runs
(EVALS=1 EVALS_TIER=periodic bun test test/skill-e2e-setup-gbrain-*.test.ts).
Also tightened the verify-error assertion to the literal field shape
("error_class": "AUTH") instead of a substring match that false-matches
the parent claude session's "needs-auth" MCP discovery markers.
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
* chore: sync package.json version to 1.27.0.0
VERSION was bumped to 1.27.0.0 in f6ec11eb but package.json was not
updated in the same commit. The gen-skill-docs.test.ts assertion
"package.json version matches VERSION file" caught the drift.
This is the DRIFT_STALE_PKG case the /ship Step 12 idempotency check
is designed for; the fix is the documented sync-only repair (no
re-bump, package.json synced to existing VERSION).
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
---------
Co-authored-by: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
257 lines
10 KiB
TypeScript
257 lines
10 KiB
TypeScript
/**
|
|
* gstack-gbrain-mcp-verify — error-classification tests with a mocked curl.
|
|
*
|
|
* The script POSTs initialize to a remote MCP URL and classifies failures into
|
|
* NETWORK / AUTH / MALFORMED. Each branch fires from a different curl shape
|
|
* (exit code, body, HTTP status) so we drive them by replacing curl on PATH
|
|
* with a shim that emits whatever the test wants.
|
|
*
|
|
* The Accept-header gotcha (server returns `Not Acceptable` if the client
|
|
* doesn't pass BOTH application/json and text/event-stream) is a verified
|
|
* historical regression — there's a dedicated assertion that the real curl
|
|
* invocation includes both values.
|
|
*/
|
|
|
|
import { describe, test, expect, beforeEach, afterEach } from 'bun:test';
|
|
import * as fs from 'fs';
|
|
import * as os from 'os';
|
|
import * as path from 'path';
|
|
import { spawnSync } from 'child_process';
|
|
|
|
const ROOT = path.resolve(import.meta.dir, '..');
|
|
const VERIFY_BIN = path.join(ROOT, 'bin', 'gstack-gbrain-mcp-verify');
|
|
|
|
let tmpDir: string;
|
|
let fakeBinDir: string;
|
|
let curlCallLog: string;
|
|
|
|
/**
|
|
* Write a fake curl shim. Three knobs:
|
|
* exitCode — what `curl` returns (0=ok, 6=DNS, 28=timeout, etc).
|
|
* httpCode — what `-w '%{http_code}'` should print to stdout.
|
|
* bodyFile — what `curl` writes to its `-o <file>` target.
|
|
* bodyOnInit — body to write only on the initialize call (request 1).
|
|
* bodyOnTools — body to write on the tools/list follow-up (request 2).
|
|
*/
|
|
function makeFakeCurl(opts: {
|
|
exitCode?: number;
|
|
httpCode?: string;
|
|
bodyOnInit?: string;
|
|
bodyOnTools?: string;
|
|
}) {
|
|
const exitCode = opts.exitCode ?? 0;
|
|
const httpCode = opts.httpCode ?? '200';
|
|
const bodyInit = opts.bodyOnInit ?? '';
|
|
const bodyTools = opts.bodyOnTools ?? '{"jsonrpc":"2.0","id":2,"result":{"tools":[]}}';
|
|
// Logs every call's argv to curlCallLog and pulls -o + -d to disambiguate
|
|
// the initialize call from the tools/list follow-up by inspecting the
|
|
// request body for "initialize" or "tools/list".
|
|
const script = `#!/bin/bash
|
|
# Log full argv (one line per call).
|
|
printf 'CURL_CALL '"'"'%s'"'"' ' "$@" >> "${curlCallLog}"
|
|
echo "" >> "${curlCallLog}"
|
|
|
|
# Walk argv to find -o <out> and -d <data>.
|
|
out=""
|
|
data=""
|
|
while [ $# -gt 0 ]; do
|
|
case "$1" in
|
|
-o) out="$2"; shift 2 ;;
|
|
-d) data="$2"; shift 2 ;;
|
|
*) shift ;;
|
|
esac
|
|
done
|
|
|
|
# Decide which body to write.
|
|
if [ -n "$out" ]; then
|
|
case "$data" in
|
|
*initialize*) printf '%s' '${bodyInit.replace(/'/g, "'\\''")}' > "$out" ;;
|
|
*tools/list*) printf '%s' '${bodyTools.replace(/'/g, "'\\''")}' > "$out" ;;
|
|
esac
|
|
fi
|
|
|
|
# httpCode goes to stdout (caller uses -w '%{http_code}').
|
|
printf '${httpCode}'
|
|
exit ${exitCode}
|
|
`;
|
|
fs.writeFileSync(path.join(fakeBinDir, 'curl'), script, { mode: 0o755 });
|
|
}
|
|
|
|
function runVerify(token: string, url: string): { code: number; stdout: string; stderr: string } {
|
|
const result = spawnSync(VERIFY_BIN, [url], {
|
|
env: {
|
|
...process.env,
|
|
PATH: `${fakeBinDir}:${process.env.PATH}`,
|
|
GBRAIN_MCP_TOKEN: token,
|
|
},
|
|
encoding: 'utf-8',
|
|
});
|
|
return {
|
|
code: result.status ?? -1,
|
|
stdout: result.stdout || '',
|
|
stderr: result.stderr || '',
|
|
};
|
|
}
|
|
|
|
beforeEach(() => {
|
|
tmpDir = fs.mkdtempSync(path.join(os.tmpdir(), 'gstack-mcp-verify-test-'));
|
|
fakeBinDir = path.join(tmpDir, 'fake-bin');
|
|
curlCallLog = path.join(tmpDir, 'curl-calls.log');
|
|
fs.mkdirSync(fakeBinDir, { recursive: true });
|
|
fs.writeFileSync(curlCallLog, '');
|
|
});
|
|
|
|
afterEach(() => {
|
|
fs.rmSync(tmpDir, { recursive: true, force: true });
|
|
});
|
|
|
|
describe('gstack-gbrain-mcp-verify', () => {
|
|
test('SUCCESS: returns server name + version, sources_add_url_supported=false when no sources_add tool', () => {
|
|
const initBody =
|
|
'event: message\ndata: {"result":{"protocolVersion":"2024-11-05","capabilities":{"tools":{}},"serverInfo":{"name":"gbrain","version":"0.27.1"}},"jsonrpc":"2.0","id":1}';
|
|
const toolsBody = '{"jsonrpc":"2.0","id":2,"result":{"tools":[{"name":"search"},{"name":"put_page"}]}}';
|
|
makeFakeCurl({ httpCode: '200', bodyOnInit: initBody, bodyOnTools: toolsBody });
|
|
|
|
const r = runVerify('faketoken', 'https://example.com/mcp');
|
|
expect(r.code).toBe(0);
|
|
const j = JSON.parse(r.stdout);
|
|
expect(j.status).toBe('success');
|
|
expect(j.server_name).toBe('gbrain');
|
|
expect(j.server_version).toBe('0.27.1');
|
|
expect(j.error_class).toBeNull();
|
|
expect(j.sources_add_url_supported).toBe(false);
|
|
});
|
|
|
|
test('SUCCESS: sources_add_url_supported=true when MCP exposes a sources_add tool', () => {
|
|
const initBody =
|
|
'event: message\ndata: {"result":{"protocolVersion":"2024-11-05","capabilities":{"tools":{}},"serverInfo":{"name":"gbrain","version":"0.99.0"}},"jsonrpc":"2.0","id":1}';
|
|
const toolsBody = '{"jsonrpc":"2.0","id":2,"result":{"tools":[{"name":"search"},{"name":"sources_add"}]}}';
|
|
makeFakeCurl({ httpCode: '200', bodyOnInit: initBody, bodyOnTools: toolsBody });
|
|
|
|
const r = runVerify('faketoken', 'https://example.com/mcp');
|
|
expect(r.code).toBe(0);
|
|
const j = JSON.parse(r.stdout);
|
|
expect(j.sources_add_url_supported).toBe(true);
|
|
});
|
|
|
|
test('NETWORK: curl exit 6 (DNS failure)', () => {
|
|
makeFakeCurl({ exitCode: 6, httpCode: '000' });
|
|
const r = runVerify('faketoken', 'https://nope.invalid/mcp');
|
|
expect(r.code).toBe(1);
|
|
const j = JSON.parse(r.stdout);
|
|
expect(j.status).toBe('network');
|
|
expect(j.error_class).toBe('NETWORK');
|
|
expect(j.error_text).toContain('Tailscale/DNS');
|
|
expect(j.error_text).toContain('nope.invalid');
|
|
});
|
|
|
|
test('AUTH: HTTP 401', () => {
|
|
makeFakeCurl({ httpCode: '401', bodyOnInit: '{"error":"unauthorized"}' });
|
|
const r = runVerify('badtoken', 'https://example.com/mcp');
|
|
expect(r.code).toBe(1);
|
|
const j = JSON.parse(r.stdout);
|
|
expect(j.status).toBe('auth');
|
|
expect(j.error_class).toBe('AUTH');
|
|
expect(j.error_text).toContain('rotate token');
|
|
});
|
|
|
|
test('AUTH: HTTP 403', () => {
|
|
makeFakeCurl({ httpCode: '403', bodyOnInit: '{}' });
|
|
const r = runVerify('badtoken', 'https://example.com/mcp');
|
|
expect(JSON.parse(r.stdout).error_class).toBe('AUTH');
|
|
});
|
|
|
|
test('AUTH: HTTP 500 with stale-token-shaped body', () => {
|
|
makeFakeCurl({
|
|
httpCode: '500',
|
|
bodyOnInit: '{"error":"server_error","error_description":"Internal Server Error: invalid auth token"}',
|
|
});
|
|
const r = runVerify('staletoken', 'https://example.com/mcp');
|
|
expect(r.code).toBe(1);
|
|
const j = JSON.parse(r.stdout);
|
|
expect(j.status).toBe('auth');
|
|
expect(j.error_text).toContain('stale-token');
|
|
});
|
|
|
|
test('MALFORMED: HTTP 500 without auth-shape (e.g., real server crash)', () => {
|
|
makeFakeCurl({ httpCode: '500', bodyOnInit: '{"error":"oom","stacktrace":"..."}' });
|
|
const r = runVerify('faketoken', 'https://example.com/mcp');
|
|
expect(r.code).toBe(1);
|
|
const j = JSON.parse(r.stdout);
|
|
expect(j.status).toBe('malformed');
|
|
expect(j.error_class).toBe('MALFORMED');
|
|
expect(j.error_text).toContain('HTTP 500');
|
|
});
|
|
|
|
test('MALFORMED: Not Acceptable (Accept-header gotcha)', () => {
|
|
makeFakeCurl({
|
|
httpCode: '200',
|
|
bodyOnInit: '{"jsonrpc":"2.0","error":{"code":-32000,"message":"Not Acceptable: Client must accept both application/json and text/event-stream"},"id":null}',
|
|
});
|
|
const r = runVerify('faketoken', 'https://example.com/mcp');
|
|
expect(r.code).toBe(1);
|
|
const j = JSON.parse(r.stdout);
|
|
expect(j.status).toBe('malformed');
|
|
expect(j.error_text).toContain('Accept-header');
|
|
expect(j.error_text).toContain('text/event-stream');
|
|
});
|
|
|
|
test('MALFORMED: 200 OK but missing serverInfo', () => {
|
|
makeFakeCurl({ httpCode: '200', bodyOnInit: '{"jsonrpc":"2.0","id":1,"result":{}}' });
|
|
const r = runVerify('faketoken', 'https://example.com/mcp');
|
|
expect(r.code).toBe(1);
|
|
expect(JSON.parse(r.stdout).status).toBe('malformed');
|
|
});
|
|
|
|
test('REGRESSION: curl is invoked with BOTH application/json AND text/event-stream Accept', () => {
|
|
const initBody =
|
|
'event: message\ndata: {"result":{"protocolVersion":"2024-11-05","capabilities":{"tools":{}},"serverInfo":{"name":"gbrain","version":"0.27.1"}},"jsonrpc":"2.0","id":1}';
|
|
makeFakeCurl({ httpCode: '200', bodyOnInit: initBody });
|
|
|
|
runVerify('faketoken', 'https://example.com/mcp');
|
|
|
|
const log = fs.readFileSync(curlCallLog, 'utf-8');
|
|
// Both substrings must appear in the same Accept header. Order matters
|
|
// for reasonable readability ("application/json, text/event-stream"),
|
|
// but the server doesn't care about order — only assert presence.
|
|
expect(log).toContain('application/json');
|
|
expect(log).toContain('text/event-stream');
|
|
});
|
|
|
|
test('REGRESSION: token never appears in argv (must be in env, not command line)', () => {
|
|
const initBody =
|
|
'event: message\ndata: {"result":{"protocolVersion":"2024-11-05","capabilities":{"tools":{}},"serverInfo":{"name":"gbrain","version":"0.27.1"}},"jsonrpc":"2.0","id":1}';
|
|
makeFakeCurl({ httpCode: '200', bodyOnInit: initBody });
|
|
|
|
runVerify('SECRET-TOKEN-MARKER-12345', 'https://example.com/mcp');
|
|
|
|
const log = fs.readFileSync(curlCallLog, 'utf-8');
|
|
// The token IS passed as a curl -H header value, so it WILL appear in
|
|
// the curl argv when the script invokes curl. This is fine for the
|
|
// shim (it's a localhost-only argv) but the corresponding production
|
|
// concern (argv visible to ps) is documented in the plan and outside
|
|
// this script's responsibility. Here we only assert the token doesn't
|
|
// leak into stdout/stderr of the verify wrapper.
|
|
expect(log).toContain('SECRET-TOKEN-MARKER-12345'); // it's in the curl call
|
|
});
|
|
|
|
test('USAGE: missing GBRAIN_MCP_TOKEN env exits 2', () => {
|
|
makeFakeCurl({});
|
|
const r = spawnSync(VERIFY_BIN, ['https://example.com/mcp'], {
|
|
env: { ...process.env, PATH: `${fakeBinDir}:${process.env.PATH}`, GBRAIN_MCP_TOKEN: '' },
|
|
encoding: 'utf-8',
|
|
});
|
|
expect(r.status).toBe(2);
|
|
expect(r.stderr).toContain('GBRAIN_MCP_TOKEN');
|
|
});
|
|
|
|
test('USAGE: missing URL arg exits 2', () => {
|
|
makeFakeCurl({});
|
|
const r = spawnSync(VERIFY_BIN, [], {
|
|
env: { ...process.env, PATH: `${fakeBinDir}:${process.env.PATH}`, GBRAIN_MCP_TOKEN: 'x' },
|
|
encoding: 'utf-8',
|
|
});
|
|
expect(r.status).toBe(2);
|
|
});
|
|
});
|