Skip to content

Piero Bosio Social Web Site Personale Logo Fediverso

Social Forum federato con il resto del mondo. Non contano le istanze, contano le persone

Designing type-safe sync/async mode support in TypeScript

Uncategorized
1 1 4
  • I recently added sync/async mode support to Optique, a type-safe CLI parser
    for TypeScript. It turned out to be one of the trickier features I've
    implemented—the object() combinator alone needed to compute a combined mode
    from all its child parsers, and TypeScript's inference kept hitting edge cases.

    What is Optique?

    Optique is a type-safe, combinatorial CLI parser for TypeScript, inspired by
    Haskell's optparse-applicative. Instead of decorators or builder patterns,
    you compose small parsers into larger ones using combinators, and TypeScript
    infers the result types.

    Here's a quick taste:

    import { object } from "@optique/core/constructs";
    import { argument, option } from "@optique/core/primitives";
    import { string, integer } from "@optique/core/valueparser";
    import { run } from "@optique/run";
    
    const cli = object({
      name: argument(string()),
      count: option("-n", "--count", integer()),
    });
    
    // TypeScript infers: { name: string; count: number | undefined }
    const result = run(cli);  // sync by default
    

    The type inference works through arbitrarily deep compositions—in most cases,
    you don't need explicit type annotations.

    How it started

    Lucas Garron (@lgarron@mastodon.social) opened an issue requesting
    async support for shell completions. He wanted to provide
    <kbd>Tab</kbd>-completion suggestions by running shell commands like
    git for-each-ref to list branches and tags.

    // Lucas's example: fetching Git branches and tags in parallel
    const [branches, tags] = await Promise.all([
      $`git for-each-ref --format='%(refname:short)' refs/heads/`.text(),
      $`git for-each-ref --format='%(refname:short)' refs/tags/`.text(),
    ]);
    

    At first, I didn't like the idea. Optique's entire API was synchronous, which
    made it simpler to reason about and avoided the “async infection” problem where
    one async function forces everything upstream to become async. I argued that
    shell completion should be near-instantaneous, and if you need async data, you
    should cache it at startup.

    But Lucas pushed back. The filesystem is a database, and many useful
    completions inherently require async work—Git refs change constantly, and
    pre-caching everything at startup doesn't scale for large repos. Fair point.

    What I needed to solve

    So, how do you support both sync and async execution modes in a composable
    parser library while maintaining type safety?

    The key requirements were:

    • parse() returns T or Promise<T>
    • complete() returns T or Promise<T>
    • suggest() returns Iterable<T> or AsyncIterable<T>
    • When combining parsers, if any parser is async, the combined result
      must be async
    • Existing sync code should continue to work unchanged

    The fourth requirement is the tricky one. Consider this:

    const syncParser = flag("--verbose");
    const asyncParser = option("--branch", asyncValueParser);
    
    // What's the type of this?
    const combined = object({ verbose: syncParser, branch: asyncParser });
    

    The combined parser should be async because one of its fields is async.
    This means we need type-level logic to compute the combined mode.

    Five design options

    I explored five different approaches, each with its own trade-offs.

    Option A: conditional types with mode parameter

    Add a mode type parameter to Parser and use conditional types:

    type Mode = "sync" | "async";
    
    type ModeValue<M extends Mode, T> = M extends "async" ? Promise<T> : T;
    
    interface Parser<M extends Mode, TValue, TState> {
      parse(context: ParserContext<TState>): ModeValue<M, ParserResult<TState>>;
      // ...
    }
    

    The challenge is computing combined modes:

    type CombineModes<T extends Record<string, Parser<any, any, any>>> =
      T[keyof T] extends Parser<infer M, any, any>
        ? M extends "async" ? "async" : "sync"
        : never;
    

    Option B: mode parameter with default value

    A variant of Option A, but place the mode parameter first with a default
    of "sync":

    interface Parser<M extends Mode = "sync", TValue, TState> {
      readonly $mode: M;
      // ...
    }
    

    The default value maintains backward compatibility—existing user code keeps
    working without changes.

    Option C: separate interfaces

    Define completely separate Parser and AsyncParser interfaces with
    explicit conversion:

    interface Parser<TValue, TState> { /* sync methods */ }
    interface AsyncParser<TValue, TState> { /* async methods */ }
    
    function toAsync<T, S>(parser: Parser<T, S>): AsyncParser<T, S>;
    

    Simpler to understand, but requires code duplication and explicit conversions.

    Option D: union return types for suggest() only

    The minimal approach. Only allow suggest() to be async:

    interface Parser<TValue, TState> {
      parse(context: ParserContext<TState>): ParserResult<TState>;  // always sync
      suggest(context: ParserContext<TState>, prefix: string):
        Iterable<Suggestion> | AsyncIterable<Suggestion>;  // can be either
    }
    

    This addresses the original use case but doesn't help if async parse() is
    ever needed.

    Option E: fp-ts style HKT simulation

    Use the technique from fp-ts to simulate Higher-Kinded Types:

    interface URItoKind<A> {
      Identity: A;
      Promise: Promise<A>;
    }
    
    type Kind<F extends keyof URItoKind<any>, A> = URItoKind<A>[F];
    
    interface Parser<F extends keyof URItoKind<any>, TValue, TState> {
      parse(context: ParserContext<TState>): Kind<F, ParserResult<TState>>;
    }
    

    The most flexible approach, but with a steep learning curve.

    Testing the idea

    Rather than commit to an approach based on theoretical analysis, I created
    a prototype to test how well TypeScript handles the type inference in practice.
    I published my findings in the GitHub issue:

    Both approaches correctly handle the “any async → all async” rule at the
    type level. (…) Complex conditional types like
    ModeValue<CombineParserModes<T>, ParserResult<TState>> sometimes require
    explicit type casting in the implementation. This only affects library
    internals. The user-facing API remains clean.

    The prototype validated that Option B (explicit mode parameter with default)
    would work. I chose it for these reasons:

    • Backward compatible: The default "sync" keeps existing code working
    • Explicit: The mode is visible in both types and runtime (via a $mode
      property)
    • Debuggable: Easy to inspect the current mode at runtime
    • Better IDE support: Type information is more predictable

    How CombineModes works

    The CombineModes type computes whether a combined parser should be sync or
    async:

    type CombineModes<T extends readonly Mode[]> = "async" extends T[number]
      ? "async"
      : "sync";
    

    This type checks if "async" is present anywhere in the tuple of modes.
    If so, the result is "async"; otherwise, it's "sync".

    For combinators like object(), I needed to extract modes from parser
    objects and combine them:

    // Extract the mode from a single parser
    type ParserMode<T> = T extends Parser<infer M, unknown, unknown> ? M : never;
    
    // Combine modes from all values in a record of parsers
    type CombineObjectModes<T extends Record<string, Parser<Mode, unknown, unknown>>> =
      CombineModes<{ [K in keyof T]: ParserMode<T[K]> }[keyof T][]>;
    

    Runtime implementation

    The type system handles compile-time safety, but the implementation also needs
    runtime logic. Each parser has a $mode property that indicates its execution
    mode:

    const syncParser = option("-n", "--name", string());
    console.log(syncParser.$mode);  // "sync"
    
    const asyncParser = option("-b", "--branch", asyncValueParser);
    console.log(asyncParser.$mode);  // "async"
    

    Combinators compute their mode at construction time:

    function object<T extends Record<string, Parser<Mode, unknown, unknown>>>(
      parsers: T
    ): Parser<CombineObjectModes<T>, ObjectValue<T>, ObjectState<T>> {
      const parserKeys = Reflect.ownKeys(parsers);
      const combinedMode: Mode = parserKeys.some(
        (k) => parsers[k as keyof T].$mode === "async"
      ) ? "async" : "sync";
    
      // ... implementation
    }
    

    Refining the API

    Lucas suggested an important refinement during our
    discussion. Instead of having run() automatically choose between sync and
    async based on the parser mode, he proposed separate functions:

    Perhaps run(…) could be automatic, and runSync(…) and runAsync(…) could
    enforce that the inferred type matches what is expected.

    So we ended up with:

    • run(): automatic based on parser mode
    • runSync(): enforces sync mode at compile time
    • runAsync(): enforces async mode at compile time
    // Automatic: returns T for sync parsers, Promise<T> for async
    const result1 = run(syncParser);  // string
    const result2 = run(asyncParser);  // Promise<string>
    
    // Explicit: compile-time enforcement
    const result3 = runSync(syncParser);  // string
    const result4 = runAsync(asyncParser);  // Promise<string>
    
    // Compile error: can't use runSync with async parser
    const result5 = runSync(asyncParser);  // Type error!
    

    I applied the same pattern to parse()/parseSync()/parseAsync() and
    suggest()/suggestSync()/suggestAsync() in the facade functions.

    Creating async value parsers

    With the new API, creating an async value parser for Git branches looks
    like this:

    import type { Suggestion } from "@optique/core/parser";
    import type { ValueParser, ValueParserResult } from "@optique/core/valueparser";
    
    function gitRef(): ValueParser<"async", string> {
      return {
        $mode: "async",
        metavar: "REF",
        parse(input: string): Promise<ValueParserResult<string>> {
          return Promise.resolve({ success: true, value: input });
        },
        format(value: string): string {
          return value;
        },
        async *suggest(prefix: string): AsyncIterable<Suggestion> {
          const { $ } = await import("bun");
          const [branches, tags] = await Promise.all([
            $`git for-each-ref --format='%(refname:short)' refs/heads/`.text(),
            $`git for-each-ref --format='%(refname:short)' refs/tags/`.text(),
          ]);
          for (const ref of [...branches.split("\n"), ...tags.split("\n")]) {
            const trimmed = ref.trim();
            if (trimmed && trimmed.startsWith(prefix)) {
              yield { kind: "literal", text: trimmed };
            }
          }
        },
      };
    }
    

    Notice that parse() returns Promise.resolve() even though it's synchronous.
    This is because the ValueParser<"async", T> type requires all methods to use
    async signatures. Lucas pointed out this is a minor ergonomic issue. If only
    suggest() needs to be async, you still have to wrap parse() in a Promise.

    I considered per-method mode granularity (e.g., ValueParser<ParseMode, SuggestMode, T>), but the implementation complexity would multiply
    substantially. For now, the workaround is simple enough:

    // Option 1: Use Promise.resolve()
    parse(input) {
      return Promise.resolve({ success: true, value: input });
    }
    
    // Option 2: Mark as async and suppress the linter
    // biome-ignore lint/suspicious/useAwait: sync implementation in async ValueParser
    async parse(input) {
      return { success: true, value: input };
    }
    

    What it cost

    Supporting dual modes added significant complexity to Optique's internals.
    Every combinator needed updates:

    • Type signatures grew more complex with mode parameters
    • Mode propagation logic had to be added to every combinator
    • Dual implementations were needed for sync and async code paths
    • Type casts were sometimes necessary in the implementation to satisfy
      TypeScript

    For example, the object() combinator went from around 100 lines to around
    250 lines. The internal implementation uses conditional logic based on the
    combined mode:

    if (combinedMode === "async") {
      return {
        $mode: "async" as M,
        // ... async implementation with Promise chains
        async parse(context) {
          // ... await each field's parse result
        },
      };
    } else {
      return {
        $mode: "sync" as M,
        // ... sync implementation
        parse(context) {
          // ... directly call each field's parse
        },
      };
    }
    

    This duplication is the cost of supporting both modes without runtime overhead
    for sync-only use cases.

    Lessons learned

    Listen to users, but validate with prototypes

    My initial instinct was to resist async support. Lucas's persistence and
    concrete examples changed my mind, but I validated the approach with a
    prototype before committing. The prototype revealed practical issues (like
    TypeScript inference limits) that pure design analysis would have missed.

    Backward compatibility is worth the complexity

    Making "sync" the default mode meant existing code continued to work
    unchanged. This was a deliberate choice. Breaking changes should require
    user action, not break silently.

    Unified mode vs per-method granularity

    I chose unified mode (all methods share the same sync/async mode) over
    per-method granularity. This means users occasionally write
    Promise.resolve() for methods that don't actually need async, but the
    alternative was multiplicative complexity in the type system.

    Designing in public

    The entire design process happened in a public GitHub issue. Lucas, Giuseppe,
    and others contributed ideas that shaped the final API. The
    runSync()/runAsync() distinction came directly from Lucas's feedback.

    Conclusion

    This was one of the more challenging features I've implemented in Optique.
    TypeScript's type system is powerful enough to encode the “any async means all
    async” rule at compile time, but getting there required careful design work and
    prototyping.

    What made it work: conditional types like ModeValue<M, T> can bridge the gap
    between sync and async worlds. You pay for it with implementation complexity,
    but the user-facing API stays clean and type-safe.

    Optique 0.9.0 with async support is currently in pre-release testing. If
    you'd like to try it, check out PR #70 or install the pre-release:

    npm  add       @optique/core@0.9.0-dev.212 @optique/run@0.9.0-dev.212
    deno add --jsr @optique/core@0.9.0-dev.212 @optique/run@0.9.0-dev.212
    

    Feedback is welcome!

  • hongminhee@hollo.socialundefined hongminhee@hollo.social shared this topic on

Gli ultimi otto messaggi ricevuti dalla Federazione
Post suggeriti
  • 0 Votes
    3 Posts
    7 Views
    @tanuki no but I'll take a look, thanks for the recommendation 🙇
  • 0 Votes
    1 Posts
    5 Views
    When I started building Fedify, an ActivityPub server framework, I ran into a problem that surprised me: I couldn't figure out how to add logging. Not because logging is hard—there are dozens of mature logging libraries for JavaScript. The problem was that they're primarily designed for applications, not for libraries that want to stay unobtrusive. I wrote about this a few months ago, and the response was modest—some interest, some skepticism, and quite a bit of debate about whether the post was AI-generated. I'll be honest: English isn't my first language, so I use LLMs to polish my writing. But the ideas and technical content are mine. Several readers wanted to see a real-world example rather than theory. The problem: existing loggers assume you're building an app Fedify helps developers build federated social applications using the ActivityPub protocol. If you've ever worked with federation, you know debugging can be painful. When an activity fails to deliver, you need to answer questions like: Did the HTTP request actually go out? Was the signature generated correctly? Did the remote server reject it? Why? Was there a problem parsing the response? These questions span multiple subsystems: HTTP handling, cryptographic signatures, JSON-LD processing, queue management, and more. Without good logging, debugging turns into guesswork. But here's the dilemma I faced as a library author: if I add verbose logging to help with debugging, I risk annoying users who don't want their console cluttered with Fedify's internal chatter. If I stay silent, users struggle to diagnose issues. I looked at the existing options. With winston or Pino, I would have to either: Configure a logger inside Fedify (imposing my choices on users), or Ask users to pass a logger instance to Fedify (adding boilerplate) There's also debug, which is designed for this use case. But it doesn't give you structured, level-based logs that ops teams expect—and it relies on environment variables, which some runtimes like Deno restrict by default for security reasons. None of these felt right. So I built LogTape—a logging library designed from the ground up for library authors. And Fedify became its first real user. The solution: hierarchical categories with zero default output The key insight was simple: a library should be able to log without producing any output unless the application developer explicitly enables it. Fedify uses LogTape's hierarchical category system to give users fine-grained control over what they see. Here's how the categories are organized: Category What it logs ["fedify"] Everything from the library ["fedify", "federation", "inbox"] Incoming activities ["fedify", "federation", "outbox"] Outgoing activities ["fedify", "federation", "http"] HTTP requests and responses ["fedify", "sig", "http"] HTTP Signature operations ["fedify", "sig", "ld"] Linked Data Signature operations ["fedify", "sig", "key"] Key generation and retrieval ["fedify", "runtime", "docloader"] JSON-LD document loading ["fedify", "webfinger", "lookup"] WebFinger resource lookups …and about a dozen more. Each category corresponds to a distinct subsystem. This means a user can configure logging like this: await configure({ sinks: { console: getConsoleSink() }, loggers: [ // Show errors from all of Fedify { category: "fedify", sinks: ["console"], lowestLevel: "error" }, // But show debug info for inbox processing specifically { category: ["fedify", "federation", "inbox"], sinks: ["console"], lowestLevel: "debug" }, ], }); When something goes wrong with incoming activities, they get detailed logs for that subsystem while keeping everything else quiet. No code changes required—just configuration. Request tracing with implicit contexts The hierarchical categories solved the filtering problem, but there was another challenge: correlating logs across async boundaries. In a federated system, a single user action might trigger a cascade of operations: fetch a remote actor, verify their signature, process the activity, fan out to followers, and so on. When something fails, you need to correlate all the log entries for that specific request. Fedify uses LogTape's implicit context feature to automatically tag every log entry with a requestId: await configure({ sinks: { file: getFileSink("fedify.jsonl", { formatter: jsonLinesFormatter }) }, loggers: [ { category: "fedify", sinks: ["file"], lowestLevel: "info" }, ], contextLocalStorage: new AsyncLocalStorage(), // Enables implicit contexts }); With this configuration, every log entry automatically includes a requestId property. When you need to debug a specific request, you can filter your logs: jq 'select(.properties.requestId == "abc-123")' fedify.jsonl And you'll see every log entry from that request—across all subsystems, all in order. No manual correlation needed. The requestId is derived from standard headers when available (X-Request-Id, Traceparent, etc.), so it integrates naturally with existing observability infrastructure. What users actually see So what does all this configuration actually mean for someone using Fedify? If a Fedify user doesn't configure LogTape at all, they see nothing. No warnings about missing configuration, no default output, and minimal performance overhead—the logging calls are essentially no-ops. For basic visibility, they can enable error-level logging for all of Fedify with three lines of configuration. When debugging a specific issue, they can enable debug-level logging for just the relevant subsystem. And if they're running in production with serious observability requirements, they can pipe structured JSON logs to their monitoring system with request correlation built in. The same library code supports all these scenarios—whether the user is running on Node.js, Deno, Bun, or edge functions, without extra polyfills or shims. The user decides what they need. Lessons learned Building Fedify with LogTape taught me a few things: Design your categories early. The hierarchical structure should reflect how users will actually want to filter logs. I organized Fedify's categories around subsystems that users might need to debug independently. Use structured logging. Properties like requestId, activityId, and actorId are far more useful than string interpolation when you need to analyze logs programmatically. Implicit contexts turned out to be more useful than I expected. Being able to correlate logs across async boundaries without passing context manually made debugging distributed operations much easier. When a user reports that activity delivery failed, I can give them a single jq command to extract everything relevant. Trust your users. Some library authors worry about exposing too much internal detail through logs. I've found the opposite—users appreciate being able to see what's happening when they need to. The key is making it opt-in. Try it yourself If you're building a library and struggling with the logging question—how much to log, how to give users control, how to avoid being noisy—I'd encourage you to look at how Fedify does it. The Fedify logging documentation explains everything in detail. And if you want to understand the philosophy behind LogTape's design, my earlier post covers that. LogTape isn't trying to replace winston or Pino for application developers who are happy with those tools. It fills a different gap: logging for libraries that want to stay out of the way until users need them. If that's what you're looking for, it might be a better fit than the usual app-centric loggers.
  • 0 Votes
    1 Posts
    11 Views
    We're thrilled to announce Optique 0.7.0, a release focused on developer experience improvements and expanding Optique's ecosystem with validation library integrations. Optique is a type-safe, combinatorial CLI argument parser for TypeScript. Unlike traditional CLI libraries that rely on configuration objects, Optique lets you compose parsers from small, reusable functions—bringing the same functional composition patterns that make Zod powerful to CLI development. If you're new to Optique, check out Why Optique? to learn how this approach unlocks possibilities that configuration-based libraries simply can't match. This release introduces automatic “Did you mean?” suggestions for typos, seamless integration with Zod and Valibot validation libraries, duplicate option name detection for catching configuration bugs early, and context-aware error messages that help users understand exactly what went wrong. “Did you mean?”: Automatic typo suggestions We've all been there: you type --verbos instead of --verbose, and the CLI responds with an unhelpful “unknown option” error. Optique 0.7.0 changes this by automatically suggesting similar options when users make typos: const parser = object({ verbose: option("-v", "--verbose"), version: option("--version"), }); // User types: --verbos (typo) const result = parse(parser, ["--verbos"]); // Error: Unexpected option or argument: --verbos. // // Did you mean one of these? // --verbose // --version The suggestion system uses Levenshtein distance to find similar names, suggesting up to 3 alternatives when the edit distance is within a reasonable threshold. Suggestions work automatically for both option names and subcommand names across all parser types—option(), flag(), command(), object(), or(), and longestMatch(). See the automatic suggestions documentation for more details. Customizing suggestions You can customize how suggestions are formatted or disable them entirely through the errors option: // Custom suggestion format for option/flag parsers const portOption = option("--port", integer(), { errors: { noMatch: (invalidOption, suggestions) => suggestions.length > 0 ? message`Unknown option ${invalidOption}. Try: ${values(suggestions)}` : message`Unknown option ${invalidOption}.` } }); // Custom suggestion format for combinators const config = object({ host: option("--host", string()), port: option("--port", integer()) }, { errors: { suggestions: (suggestions) => suggestions.length > 0 ? message`Available options: ${values(suggestions)}` : [] } }); Zod and Valibot integrations Two new packages join the Optique family, bringing powerful validation capabilities from the TypeScript ecosystem to your CLI parsers. @optique/zod The new @optique/zod package lets you use Zod schemas directly as value parsers: import { option, object } from "@optique/core"; import { zod } from "@optique/zod"; import { z } from "zod"; const parser = object({ email: option("--email", zod(z.string().email())), port: option("--port", zod(z.coerce.number().int().min(1).max(65535))), format: option("--format", zod(z.enum(["json", "yaml", "xml"]))), }); The package supports both Zod v3.25.0+ and v4.0.0+, with automatic error formatting that integrates seamlessly with Optique's message system. See the Zod integration guide for complete usage examples. @optique/valibot For those who prefer a lighter bundle, @optique/valibot integrates with Valibot—a validation library with a significantly smaller footprint (~10KB vs Zod's ~52KB): import { option, object } from "@optique/core"; import { valibot } from "@optique/valibot"; import * as v from "valibot"; const parser = object({ email: option("--email", valibot(v.pipe(v.string(), v.email()))), port: option("--port", valibot(v.pipe( v.string(), v.transform(Number), v.integer(), v.minValue(1), v.maxValue(65535) ))), }); Both packages support custom error messages through their respective error handler options (zodError and valibotError), giving you full control over how validation failures are presented to users. See the Valibot integration guide for complete usage examples. Duplicate option name detection A common source of bugs in CLI applications is accidentally using the same option name in multiple places. Previously, this would silently cause ambiguous parsing where the first matching parser consumed the option. Optique 0.7.0 now validates option names at parse time and fails with a clear error message when duplicates are detected: const parser = object({ input: option("-i", "--input", string()), interactive: option("-i", "--interactive"), // Oops! -i is already used }); // Error: Duplicate option name -i found in fields: input, interactive. // Each option name must be unique within a parser combinator. This validation applies to object(), tuple(), merge(), and group() combinators. The or() combinator continues to allow duplicate option names since its branches are mutually exclusive. See the duplicate detection documentation for more details. If you have a legitimate use case for duplicate option names, you can opt out with allowDuplicates: true: const parser = object({ input: option("-i", "--input", string()), interactive: option("-i", "--interactive"), }, { allowDuplicates: true }); Context-aware error messages Error messages from combinators are now smarter about what they report. Instead of generic "No matching option or command found" messages, Optique now analyzes what the parser expects and provides specific feedback: // When only arguments are expected const parser1 = or(argument(string()), argument(integer())); // Error: Missing required argument. // When only commands are expected const parser2 = or(command("add", addParser), command("remove", removeParser)); // Error: No matching command found. // When both options and arguments are expected const parser3 = object({ port: option("--port", integer()), file: argument(string()), }); // Error: No matching option or argument found. Dynamic error messages with NoMatchContext For applications that need internationalization or context-specific messaging, the errors.noMatch option now accepts a function that receives a NoMatchContext object: const parser = or( command("add", addParser), command("remove", removeParser), { errors: { noMatch: ({ hasOptions, hasCommands, hasArguments }) => { if (hasCommands && !hasOptions && !hasArguments) { return message`일치하는 명령을 찾을 수 없습니다.`; // Korean } return message`잘못된 입력입니다.`; } } } ); Shell completion naming conventions The run() function now supports configuring whether shell completions use singular or plural naming conventions: run(parser, { completion: { name: "plural", // Uses "completions" and "--completions" } }); // Or for singular only run(parser, { completion: { name: "singular", // Uses "completion" and "--completion" } }); The default "both" accepts either form, maintaining backward compatibility while letting you enforce a consistent style in your CLI. Additional improvements Line break handling: formatMessage() now distinguishes between soft breaks (single \n, converted to spaces) and hard breaks (double \n\n, creating paragraph separations), improving multi-line error message formatting. New utility functions: Added extractOptionNames() and extractArgumentMetavars() to the @optique/core/usage module for programmatic access to parser metadata. Installation deno add --jsr @optique/core @optique/run npm add @optique/core @optique/run pnpm add @optique/core @optique/run yarn add @optique/core @optique/run bun add @optique/core @optique/run For validation library integrations: # Zod integration deno add jsr:@optique/zod # Deno npm add @optique/zod # npm/pnpm/yarn/bun # Valibot integration deno add jsr:@optique/valibot # Deno npm add @optique/valibot # npm/pnpm/yarn/bun Looking forward This release represents our commitment to making CLI development in TypeScript as smooth as possible. The “Did you mean?” suggestions and validation library integrations were among the most requested features, and we're excited to see how they improve your CLI applications. For detailed documentation and examples, visit the Optique documentation. We welcome your feedback and contributions on GitHub!
  • 0 Votes
    1 Posts
    11 Views
    Optique 0.6.0 is here, bringing intelligent shell completion to your type-safe command-line applications. This release introduces built-in completion support for Bash, zsh, fish, PowerShell, and Nushell, making your CLIs more discoverable and user-friendly—all without sacrificing type safety or requiring duplicate definitions. For those new to [Optique]: it's a TypeScript CLI parser library that takes a fundamentally different approach from traditional configuration-based parsers. Instead of describing your CLI with configuration objects, you compose parsers from small, type-safe functions. TypeScript automatically infers the exact types of your parsed data, ensuring compile-time safety while the parser structure itself provides runtime validation. Think of it as bringing the composability of parser combinators (inspired by Haskell's optparse-applicative) together with the type safety of TypeScript's type system. Shell completion that just works The standout feature of this release is comprehensive shell completion support. Unlike many CLI frameworks that require separate completion definitions, Optique's completion system leverages the same parser structure used for argument parsing. This means your completion suggestions automatically stay synchronized with your CLI's actual behavior—no duplicate definitions, no manual maintenance. import { object } from "@optique/core/constructs"; import { argument, option } from "@optique/core/primitives"; import { string, choice } from "@optique/core/valueparser"; import { run } from "@optique/run"; const parser = object({ format: option("-f", "--format", choice(["json", "yaml", "xml"])), output: option("-o", "--output", string({ metavar: "FILE" })), verbose: option("-v", "--verbose"), input: argument(string({ metavar: "INPUT" })), }); // Enable completion with a single option const config = run(parser, { completion: "both" }); Users can now press Tab to get intelligent suggestions: myapp <TAB> # Shows available commands and options myapp --format <TAB> # Shows: json, yaml, xml myapp --format=<TAB> # Same suggestions with equals syntax myapp -<TAB> # Shows: -f, -o, -v, and other short options Setting up completion is straightforward. Users generate a completion script for their shell and source it: # Bash myapp completion bash > ~/.bashrc.d/myapp.bash source ~/.bashrc.d/myapp.bash # zsh myapp completion zsh > ~/.zsh/completions/_myapp # fish myapp completion fish > ~/.config/fish/completions/myapp.fish # PowerShell myapp completion pwsh > myapp-completion.ps1 . ./myapp-completion.ps1 # Nushell myapp completion nu | save myapp-completion.nu source myapp-completion.nu The completion system works automatically with all Optique parser types. When you use choice() value parsers, the available options become completion suggestions. When you use path() parsers, file system completion kicks in with proper handling of extensions and file types. Subcommands, options, and arguments all provide context-aware suggestions. What makes Optique's completion special is that it leverages the same parser structure used for argument parsing. Every parser has an optional suggest() method that provides context-aware suggestions based on the current input. Parser combinators like object() and or() automatically aggregate suggestions from their constituent parsers, ensuring your completion logic stays in your TypeScript code where it benefits from type safety and testing. Optique handles the differences between shells transparently. Bash uses the complete command with proper handling of word splitting, zsh leverages its powerful compdef system with completion descriptions, fish provides tab-separated format with automatic file type detection, PowerShell uses Register-ArgumentCompleter with AST-based parsing, and Nushell integrates with its external completer system. For file and directory completions, Optique delegates to each shell's native file completion system, ensuring proper handling of spaces, symlinks, and platform-specific path conventions. Custom completion suggestions For domain-specific value parsers, you can implement custom completion logic that provides intelligent suggestions based on your application's needs: import type { ValueParser, ValueParserResult } from "@optique/core/valueparser"; import type { Suggestion } from "@optique/core/parser"; import { message } from "@optique/core/message"; function httpMethod(): ValueParser<string> { const methods = ["GET", "POST", "PUT", "DELETE", "PATCH", "HEAD", "OPTIONS"]; return { metavar: "METHOD", parse(input: string): ValueParserResult<string> { const method = input.toUpperCase(); if (methods.includes(method)) { return { success: true, value: method }; } return { success: false, error: message`Invalid HTTP method: ${input}. Valid methods: ${methods.join(", ")}.`, }; }, format(value: string): string { return value; }, *suggest(prefix: string): Iterable<Suggestion> { for (const method of methods) { if (method.toLowerCase().startsWith(prefix.toLowerCase())) { yield { kind: "literal", text: method, description: message`HTTP ${method} request method` }; } } }, }; } The built-in value parsers also provide intelligent suggestions. For instance, the locale() parser suggests common locale identifiers, the url() parser offers protocol completions when configured with allowedProtocols, and the timezone parsers from @optique/temporal use Intl.supportedValuesOf() for dynamic timezone suggestions. Enhanced command documentation This release also introduces new documentation capabilities for the command() parser. You can now provide separate brief and description texts, along with a footer for examples and additional information: import { command, object, constant } from "@optique/core/primitives"; import { message } from "@optique/core/message"; const deployCommand = command( "deploy", object({ action: constant("deploy"), // ... options }), { brief: message`Deploy application to production`, // Shown in command list description: message`Deploy the application to the production environment. This command handles database migrations, asset compilation, and cache warming automatically. It performs health checks before switching traffic to ensure zero-downtime deployment.`, // Shown in detailed help footer: message`Examples: myapp deploy --environment staging --dry-run myapp deploy --environment production --force For deployment documentation, see: https://docs.example.com/deploy` } ); The brief text appears when listing commands (like myapp help), while description provides detailed information when viewing command-specific help (myapp deploy --help or myapp help deploy). The footer appears at the bottom of the help text, perfect for examples and additional resources. Command-line example formatting To make help text and examples clearer, we've added a new commandLine() message term type. This displays command-line snippets with distinct cyan coloring in terminals, making it immediately clear what users should type: import { message, commandLine } from "@optique/core/message"; import { run } from "@optique/run"; const config = run(parser, { footer: message`Examples: ${commandLine("myapp --format json input.txt")} ${commandLine("myapp --format=yaml --output result.yml data.txt")} To enable shell completion: ${commandLine("myapp completion bash > ~/.bashrc.d/myapp.bash")} ${commandLine("source ~/.bashrc.d/myapp.bash")}`, completion: "both" }); These command examples stand out visually in help text, making it easier for users to understand how to use your CLI. Migration guide If you're already using Optique, adding completion support is straightforward: Update to Optique 0.6.0 Add the completion option to your run() configuration: // Before const config = run(parser, { help: "both" }); // After const config = run(parser, { help: "both", completion: "both" // Adds both 'completion' command and '--completion' option }); That's it! Your CLI now supports shell completion. The completion option accepts three modes: "command": Only the completion subcommand (e.g., myapp completion bash) "option": Only the --completion option (e.g., myapp --completion bash) "both": Both patterns work For custom value parsers, you can optionally add a suggest() method to provide domain-specific completions. Existing parsers continue to work without modification—they just won't provide custom suggestions beyond what the parser structure implies. Looking forward Shell completion has been one of the most requested features for Optique, and we're thrilled to deliver it in a way that maintains our core principles: type safety, composability, and zero duplication. Your parser definitions remain the single source of truth for both parsing and completion behavior. This release represents a significant step toward making Optique-based CLIs as user-friendly as they are developer-friendly. The completion system proves that we can provide sophisticated runtime features without sacrificing the compile-time guarantees that make Optique unique. We hope you find the new shell completion feature useful and look forward to seeing what you build with it! Getting started To start using Optique 0.6.0: deno add --jsr @optique/core@^0.6.0 @optique/run@^0.6.0 npm add @optique/core@^0.6.0 @optique/run@^0.6.0 pnpm add @optique/core@^0.6.0 @optique/run@^0.6.0 yarn add @optique/core@^0.6.0 @optique/run@^0.6.0 bun add @optique/core@^0.6.0 @optique/run@^0.6.0 For complete documentation, visit optique.dev. Check out the new shell completion guide for detailed setup instructions and advanced usage patterns. For bug reports and feature requests, please visit our GitHub repository.