From 26b7ddf56a90e54fa5525c5f11ba161e331a0a07 Mon Sep 17 00:00:00 2001 From: Johnny Graettinger Date: Tue, 14 Nov 2023 21:32:59 +0000 Subject: [PATCH] go.mod: bump gazette version and regenerate --- client/dist/src/gen/broker/protocol/broker.ts | 1137 ++++++++-------- .../src/gen/consumer/protocol/consumer.ts | 1144 +++++++++-------- .../types/gen/broker/protocol/broker.d.ts | 265 ++-- .../types/gen/consumer/protocol/consumer.d.ts | 240 ++-- client/src/gen/broker/protocol/broker.ts | 1137 ++++++++-------- client/src/gen/consumer/protocol/consumer.ts | 1144 +++++++++-------- gen/broker/protocol/protocol.swagger.json | 5 + go.mod | 4 +- go.sum | 264 +--- 9 files changed, 2610 insertions(+), 2730 deletions(-) diff --git a/client/dist/src/gen/broker/protocol/broker.ts b/client/dist/src/gen/broker/protocol/broker.ts index b2366bc..964b6ea 100644 --- a/client/dist/src/gen/broker/protocol/broker.ts +++ b/client/dist/src/gen/broker/protocol/broker.ts @@ -10,644 +10,675 @@ */ /** - * Change defines an insertion, update, or deletion to be applied to the set - * of JournalSpecs. Exactly one of |upsert| or |delete| must be set. - */ +* Change defines an insertion, update, or deletion to be applied to the set +of JournalSpecs. Exactly one of |upsert| or |delete| must be set. +*/ export interface ApplyRequestChange { - /** - * Expected ModRevision of the current JournalSpec. If the Journal is being - * created, expect_mod_revision is zero. - * @format int64 - */ - expectModRevision?: string; - /** - * JournalSpec to be updated (if expect_mod_revision > 0) or created - * (if expect_mod_revision == 0). - */ - upsert?: ProtocolJournalSpec; - /** Journal to be deleted. expect_mod_revision must not be zero. */ - delete?: string; + /** + * Expected ModRevision of the current JournalSpec. If the Journal is being + * created, expect_mod_revision is zero. + * @format int64 + */ + expectModRevision?: string; + /** + * JournalSpec to be updated (if expect_mod_revision > 0) or created + * (if expect_mod_revision == 0). + */ + upsert?: ProtocolJournalSpec; + /** Journal to be deleted. expect_mod_revision must not be zero. */ + delete?: string; } -/** Fragments of the Response. */ +/** + * Fragments of the Response. + */ export interface FragmentsResponseFragment { - /** - * Fragment is a content-addressed description of a contiguous Journal span, - * defined by the [begin, end) offset range covered by the Fragment and the - * SHA1 sum of the corresponding Journal content. - */ - spec?: ProtocolFragment; - /** - * SignedURL is a temporary URL at which a direct GET of the Fragment may - * be issued, signed by the broker's credentials. Set only if the request - * specified a SignatureTTL. - */ - signedUrl?: string; + /** + * Fragment is a content-addressed description of a contiguous Journal span, + * defined by the [begin, end) offset range covered by the Fragment and the + * SHA1 sum of the corresponding Journal content. + */ + spec?: ProtocolFragment; + /** + * SignedURL is a temporary URL at which a direct GET of the Fragment may + * be issued, signed by the broker's credentials. Set only if the request + * specified a SignatureTTL. + */ + signedUrl?: string; } /** - * Etcd represents the effective Etcd MVCC state under which a Gazette broker - * is operating in its processing of requests and responses. Its inclusion - * allows brokers to reason about relative "happened before" Revision ordering - * of apparent routing conflicts in proxied or replicated requests, as well - * as enabling sanity checks over equality of Etcd ClusterId (and precluding, - * for example, split-brain scenarios where different brokers are backed by - * different Etcd clusters). Etcd is kept in sync with - * etcdserverpb.ResponseHeader. - */ +* Etcd represents the effective Etcd MVCC state under which a Gazette broker +is operating in its processing of requests and responses. Its inclusion +allows brokers to reason about relative "happened before" Revision ordering +of apparent routing conflicts in proxied or replicated requests, as well +as enabling sanity checks over equality of Etcd ClusterId (and precluding, +for example, split-brain scenarios where different brokers are backed by +different Etcd clusters). Etcd is kept in sync with +etcdserverpb.ResponseHeader. +*/ export interface HeaderEtcd { - /** - * cluster_id is the ID of the cluster. - * @format uint64 - */ - clusterId?: string; - /** - * member_id is the ID of the member. - * @format uint64 - */ - memberId?: string; - /** - * revision is the Etcd key-value store revision when the request was - * applied. - * @format int64 - */ - revision?: string; - /** - * raft_term is the raft term when the request was applied. - * @format uint64 - */ - raftTerm?: string; + /** + * cluster_id is the ID of the cluster. + * @format uint64 + */ + clusterId?: string; + /** + * member_id is the ID of the member. + * @format uint64 + */ + memberId?: string; + /** + * revision is the Etcd key-value store revision when the request was + * applied. + * @format int64 + */ + revision?: string; + /** + * raft_term is the raft term when the request was applied. + * @format uint64 + */ + raftTerm?: string; } -/** Journals of the response. */ +/** + * Journals of the response. + */ export interface ListResponseJournal { - /** JournalSpec describes a Journal and its configuration. */ - spec?: ProtocolJournalSpec; - /** - * Current ModRevision of the JournalSpec. - * @format int64 - */ - modRevision?: string; - /** Route of the journal, including endpoints. */ - route?: ProtocolRoute; + /** JournalSpec describes a Journal and its configuration. */ + spec?: ProtocolJournalSpec; + /** + * Current ModRevision of the JournalSpec. + * @format int64 + */ + modRevision?: string; + /** Route of the journal, including endpoints. */ + route?: ProtocolRoute; } -/** ID composes a zone and a suffix to uniquely identify a ProcessSpec. */ +/** + * ID composes a zone and a suffix to uniquely identify a ProcessSpec. + */ export interface ProcessSpecID { - /** - * "Zone" in which the process is running. Zones may be AWS, Azure, or - * Google Cloud Platform zone identifiers, or rack locations within a colo, - * or given some other custom meaning. Gazette will replicate across - * multiple zones, and seeks to minimize traffic which must cross zones (for - * example, by proxying reads to a broker in the current zone). - */ - zone?: string; - /** - * Unique suffix of the process within |zone|. It is permissible for a - * suffix value to repeat across zones, but never within zones. In practice, - * it's recommended to use a FQDN, Kubernetes Pod name, or comparable unique - * and self-describing value as the ID suffix. - */ - suffix?: string; + /** + * "Zone" in which the process is running. Zones may be AWS, Azure, or + * Google Cloud Platform zone identifiers, or rack locations within a colo, + * or given some other custom meaning. Gazette will replicate across + * multiple zones, and seeks to minimize traffic which must cross zones (for + * example, by proxying reads to a broker in the current zone). + */ + zone?: string; + /** + * Unique suffix of the process within |zone|. It is permissible for a + * suffix value to repeat across zones, but never within zones. In practice, + * it's recommended to use a FQDN, Kubernetes Pod name, or comparable unique + * and self-describing value as the ID suffix. + */ + suffix?: string; } export interface ProtobufAny { - typeUrl?: string; - /** @format byte */ - value?: string; + typeUrl?: string; + /** @format byte */ + value?: string; } -/** AppendResponse is the unary response message of the broker Append RPC. */ +/** + * AppendResponse is the unary response message of the broker Append RPC. + */ export interface ProtocolAppendResponse { - /** Status of the Append RPC. */ - status?: ProtocolStatus; - /** Header of the response. */ - header?: ProtocolHeader; - /** - * If status is OK, then |commit| is the Fragment which places the - * committed Append content within the Journal. - */ - commit?: ProtocolFragment; - /** Current registers of the journal. */ - registers?: ProtocolLabelSet; - /** - * Total number of RPC content chunks processed in this append. - * @format int64 - */ - totalChunks?: string; - /** - * Number of content chunks which were delayed by journal flow control. - * @format int64 - */ - delayedChunks?: string; + /** Status of the Append RPC. */ + status?: ProtocolStatus; + /** Header of the response. */ + header?: ProtocolHeader; + /** + * If status is OK, then |commit| is the Fragment which places the + * committed Append content within the Journal. + */ + commit?: ProtocolFragment; + /** Current registers of the journal. */ + registers?: ProtocolLabelSet; + /** + * Total number of RPC content chunks processed in this append. + * @format int64 + */ + totalChunks?: string; + /** + * Number of content chunks which were delayed by journal flow control. + * @format int64 + */ + delayedChunks?: string; } -/** ApplyResponse is the unary response message of the broker Apply RPC. */ +/** + * ApplyResponse is the unary response message of the broker Apply RPC. + */ export interface ProtocolApplyResponse { - /** Status of the Apply RPC. */ - status?: ProtocolStatus; - /** Header of the response. */ - header?: ProtocolHeader; + /** Status of the Apply RPC. */ + status?: ProtocolStatus; + /** Header of the response. */ + header?: ProtocolHeader; } /** - * CompressionCode defines codecs known to Gazette. - * - * - INVALID: INVALID is the zero-valued CompressionCodec, and is not a valid codec. - * - NONE: NONE encodes Fragments without any applied compression, with default suffix - * ".raw". - * - GZIP: GZIP encodes Fragments using the Gzip library, with default suffix ".gz". - * - ZSTANDARD: ZSTANDARD encodes Fragments using the ZStandard library, with default - * suffix ".zst". - * - SNAPPY: SNAPPY encodes Fragments using the Snappy library, with default suffix - * ".sz". - * - GZIP_OFFLOAD_DECOMPRESSION: GZIP_OFFLOAD_DECOMPRESSION is the GZIP codec with additional behavior - * around reads and writes to remote Fragment stores, designed to offload - * the work of decompression onto compatible stores. Specifically: - * * Fragments are written with a "Content-Encoding: gzip" header. - * * Client read requests are made with "Accept-Encoding: identity". - * This can be helpful in contexts where reader IO bandwidth to the storage - * API is unconstrained, as the cost of decompression is offloaded to the - * store and CPU-intensive batch readers may receive a parallelism benefit. - * While this codec may provide substantial read-time performance - * improvements, it is an advanced configuration and the "Content-Encoding" - * header handling can be subtle and sometimes confusing. It uses the default - * suffix ".gzod". - * @default "INVALID" - */ +* CompressionCode defines codecs known to Gazette. + + - INVALID: INVALID is the zero-valued CompressionCodec, and is not a valid codec. + - NONE: NONE encodes Fragments without any applied compression, with default suffix +".raw". + - GZIP: GZIP encodes Fragments using the Gzip library, with default suffix ".gz". + - ZSTANDARD: ZSTANDARD encodes Fragments using the ZStandard library, with default +suffix ".zst". + - SNAPPY: SNAPPY encodes Fragments using the Snappy library, with default suffix +".sz". + - GZIP_OFFLOAD_DECOMPRESSION: GZIP_OFFLOAD_DECOMPRESSION is the GZIP codec with additional behavior +around reads and writes to remote Fragment stores, designed to offload +the work of decompression onto compatible stores. Specifically: + * Fragments are written with a "Content-Encoding: gzip" header. + * Client read requests are made with "Accept-Encoding: identity". +This can be helpful in contexts where reader IO bandwidth to the storage +API is unconstrained, as the cost of decompression is offloaded to the +store and CPU-intensive batch readers may receive a parallelism benefit. +While this codec may provide substantial read-time performance +improvements, it is an advanced configuration and the "Content-Encoding" +header handling can be subtle and sometimes confusing. It uses the default +suffix ".gzod". +*/ export type ProtocolCompressionCodec = - | "INVALID" - | "NONE" - | "GZIP" - | "ZSTANDARD" - | "SNAPPY" - | "GZIP_OFFLOAD_DECOMPRESSION"; + | 'INVALID' + | 'NONE' + | 'GZIP' + | 'ZSTANDARD' + | 'SNAPPY' + | 'GZIP_OFFLOAD_DECOMPRESSION'; /** - * Fragment is a content-addressed description of a contiguous Journal span, - * defined by the [begin, end) offset range covered by the Fragment and the - * SHA1 sum of the corresponding Journal content. - */ +* Fragment is a content-addressed description of a contiguous Journal span, +defined by the [begin, end) offset range covered by the Fragment and the +SHA1 sum of the corresponding Journal content. +*/ export interface ProtocolFragment { - /** Journal of the Fragment. */ - journal?: string; - /** - * Begin (inclusive) and end (exclusive) offset of the Fragment within the - * Journal. - * @format int64 - */ - begin?: string; - /** @format int64 */ - end?: string; - /** SHA1 sum of the Fragment's content. */ - sum?: ProtocolSHA1Sum; - /** Codec with which the Fragment's content is compressed. */ - compressionCodec?: ProtocolCompressionCodec; - /** - * Fragment store which backs the Fragment. Empty if the Fragment has yet to - * be persisted and is still local to a Broker. - */ - backingStore?: string; - /** - * Modification timestamp of the Fragment within the backing store, - * represented as seconds since the epoch. - * @format int64 - */ - modTime?: string; - /** - * Path postfix under which the fragment is persisted to the store. - * The complete Fragment store path is built from any path components of the - * backing store, followed by the journal name, followed by the path postfix. - */ - pathPostfix?: string; + /** Journal of the Fragment. */ + journal?: string; + /** + * Begin (inclusive) and end (exclusive) offset of the Fragment within the + * Journal. + * @format int64 + */ + begin?: string; + /** @format int64 */ + end?: string; + /** SHA1 sum of the Fragment's content. */ + sum?: ProtocolSHA1Sum; + /** Codec with which the Fragment's content is compressed. */ + compressionCodec?: ProtocolCompressionCodec; + /** + * Fragment store which backs the Fragment. Empty if the Fragment has yet to + * be persisted and is still local to a Broker. + */ + backingStore?: string; + /** + * Modification timestamp of the Fragment within the backing store, + * represented as seconds since the epoch. + * @format int64 + */ + modTime?: string; + /** + * Path postfix under which the fragment is persisted to the store. + * The complete Fragment store path is built from any path components of the + * backing store, followed by the journal name, followed by the path postfix. + */ + pathPostfix?: string; } /** - * FragmentsResponse is the unary response message of the broker ListFragments - * RPC. - */ +* FragmentsResponse is the unary response message of the broker ListFragments +RPC. +*/ export interface ProtocolFragmentsResponse { - /** Status of the Apply RPC. */ - status?: ProtocolStatus; - /** Header of the response. */ - header?: ProtocolHeader; - fragments?: FragmentsResponseFragment[]; - /** - * The NextPageToke value to be returned on subsequent Fragments requests. If - * the value is zero then there are no more fragments to be returned for this - * page. - * @format int64 - */ - nextPageToken?: string; + /** Status of the Apply RPC. */ + status?: ProtocolStatus; + /** Header of the response. */ + header?: ProtocolHeader; + fragments?: FragmentsResponseFragment[]; + /** + * The NextPageToke value to be returned on subsequent Fragments requests. If + * the value is zero then there are no more fragments to be returned for this + * page. + * @format int64 + */ + nextPageToken?: string; } /** - * Header captures metadata such as the process responsible for processing - * an RPC, and its effective Etcd state. - */ +* Header captures metadata such as the process responsible for processing +an RPC, and its effective Etcd state. +*/ export interface ProtocolHeader { - /** - * ID of the process responsible for request processing. May be empty iff - * Header is being used within a proxied request, and that request may be - * dispatched to any member of the Route. - */ - processId?: ProcessSpecID; - /** - * Route of processes specifically responsible for this RPC, or an empty Route - * if any process is capable of serving the RPC. - */ - route?: ProtocolRoute; - /** - * Etcd represents the effective Etcd MVCC state under which a Gazette broker - * is operating in its processing of requests and responses. Its inclusion - * allows brokers to reason about relative "happened before" Revision ordering - * of apparent routing conflicts in proxied or replicated requests, as well - * as enabling sanity checks over equality of Etcd ClusterId (and precluding, - * for example, split-brain scenarios where different brokers are backed by - * different Etcd clusters). Etcd is kept in sync with - * etcdserverpb.ResponseHeader. - */ - etcd?: HeaderEtcd; + /** + * ID of the process responsible for request processing. May be empty iff + * Header is being used within a proxied request, and that request may be + * dispatched to any member of the Route. + */ + processId?: ProcessSpecID; + /** + * Route of processes specifically responsible for this RPC, or an empty Route + * if any process is capable of serving the RPC. + */ + route?: ProtocolRoute; + /** + * Etcd represents the effective Etcd MVCC state under which a Gazette broker + * is operating in its processing of requests and responses. Its inclusion + * allows brokers to reason about relative "happened before" Revision ordering + * of apparent routing conflicts in proxied or replicated requests, as well + * as enabling sanity checks over equality of Etcd ClusterId (and precluding, + * for example, split-brain scenarios where different brokers are backed by + * different Etcd clusters). Etcd is kept in sync with + * etcdserverpb.ResponseHeader. + */ + etcd?: HeaderEtcd; } -/** JournalSpec describes a Journal and its configuration. */ +/** + * JournalSpec describes a Journal and its configuration. + */ export interface ProtocolJournalSpec { - /** Name of the Journal. */ - name?: string; - /** - * Desired replication of this Journal. This defines the Journal's tolerance - * to broker failures before data loss can occur (eg, a replication factor - * of three means two failures are tolerated). - * @format int32 - */ - replication?: number; - /** - * User-defined Labels of this JournalSpec. Two label names are reserved - * and may not be used within a JournalSpec's Labels: "name" and "prefix". - */ - labels?: ProtocolLabelSet; - /** - * Fragment is JournalSpec configuration which pertains to the creation, - * persistence, and indexing of the Journal's Fragments. - */ - fragment?: ProtocolJournalSpecFragment; - /** - * Flags of the Journal, as a combination of Flag enum values. The Flag enum - * is not used directly, as protobuf enums do not allow for or'ed bitfields. - * @format int64 - */ - flags?: number; - /** - * Maximum rate, in bytes-per-second, at which appends of this journal will - * be processed. If zero (the default), no rate limiting is applied. A global - * rate limit still may be in effect, in which case the effective rate is the - * smaller of the journal vs global rate. - * @format int64 - */ - maxAppendRate?: string; + /** Name of the Journal. */ + name?: string; + /** + * Desired replication of this Journal. This defines the Journal's tolerance + * to broker failures before data loss can occur (eg, a replication factor + * of three means two failures are tolerated). + * @format int32 + */ + replication?: number; + /** + * User-defined Labels of this JournalSpec. Two label names are reserved + * and may not be used within a JournalSpec's Labels: "name" and "prefix". + */ + labels?: ProtocolLabelSet; + /** + * Fragment is JournalSpec configuration which pertains to the creation, + * persistence, and indexing of the Journal's Fragments. + */ + fragment?: ProtocolJournalSpecFragment; + /** + * Flags of the Journal, as a combination of Flag enum values. The Flag enum + * is not used directly, as protobuf enums do not allow for or'ed bitfields. + * @format int64 + */ + flags?: number; + /** + * Maximum rate, in bytes-per-second, at which appends of this journal will + * be processed. If zero (the default), no rate limiting is applied. A global + * rate limit still may be in effect, in which case the effective rate is the + * smaller of the journal vs global rate. + * @format int64 + */ + maxAppendRate?: string; } /** - * Fragment is JournalSpec configuration which pertains to the creation, - * persistence, and indexing of the Journal's Fragments. - */ +* Fragment is JournalSpec configuration which pertains to the creation, +persistence, and indexing of the Journal's Fragments. +*/ export interface ProtocolJournalSpecFragment { - /** - * Target content length of each Fragment. In normal operation after - * Fragments reach at least this length, they will be closed and new ones - * begun. Note lengths may be smaller at times (eg, due to changes in - * Journal routing topology). Content length differs from Fragment file - * size, in that the former reflects uncompressed bytes. - * @format int64 - */ - length?: string; - /** Codec used to compress Journal Fragments. */ - compressionCodec?: ProtocolCompressionCodec; - /** - * Storage backend base path for this Journal's Fragments. Must be in URL - * form, with the choice of backend defined by the scheme. The full path of - * a Journal's Fragment is derived by joining the store path with the - * Fragment's ContentPath. Eg, given a fragment_store of - * "s3://My-AWS-bucket/a/prefix" and a JournalSpec of name "my/journal", - * a complete Fragment path might be: - * "s3://My-AWS-bucket/a/prefix/my/journal/000123-000456-789abcdef.gzip - * Multiple stores may be specified, in which case the Journal's Fragments - * are the union of all Fragments present across all stores, and new - * Fragments always persist to the first specified store. This can be - * helpful in performing incremental migrations, where new Journal content - * is written to the new store, while content in the old store remains - * available (and, depending on fragment_retention or recovery log pruning, - * may eventually be removed). - * - * If no stores are specified, the Journal is still use-able but will - * not persist Fragments to any a backing fragment store. This allows for - * real-time streaming use cases where reads of historical data are not - * needed. - */ - stores?: string[]; - /** - * Interval of time between refreshes of remote Fragment listings from - * configured fragment_stores. - */ - refreshInterval?: string; - /** - * Retention duration for historical Fragments of this Journal within the - * Fragment stores. If less than or equal to zero, Fragments are retained - * indefinitely. - */ - retention?: string; - /** - * Flush interval defines a uniform UTC time segment which, when passed, - * will prompt brokers to close and persist a fragment presently being - * written. - * - * Flush interval may be helpful in integrating the journal with a regularly - * scheduled batch work-flow which processes new files from the fragment - * store and has no particular awareness of Gazette. For example, setting - * flush_interval to 3600s will cause brokers to persist their present - * fragment on the hour, every hour, even if it has not yet reached its - * target length. A batch work-flow running at 5 minutes past the hour is - * then reasonably assured of seeing all events from the past hour. - * - * See also "gazctl journals fragments --help" for more discussion. - */ - flushInterval?: string; - /** - * Path postfix template is a Go template which evaluates to a partial - * path under which fragments are persisted to the store. A complete - * fragment path is constructed by appending path components from the - * fragment store, then the journal name, and then the postfix template. - * Path post-fixes can help in maintaining Hive compatible partitioning - * over fragment creation time. The fields ".Spool" and ".JournalSpec" - * are available for introspection in the template. For example, - * to partition on the UTC date and hour of creation, use: - * date={{ .Spool.FirstAppendTime.Format "2006-01-02" }}/hour={{ - * .Spool.FirstAppendTime.Format "15" }} - * - * Which will produce a path postfix like "date=2019-11-19/hour=22". - */ - pathPostfixTemplate?: string; + /** + * Target content length of each Fragment. In normal operation after + * Fragments reach at least this length, they will be closed and new ones + * begun. Note lengths may be smaller at times (eg, due to changes in + * Journal routing topology). Content length differs from Fragment file + * size, in that the former reflects uncompressed bytes. + * @format int64 + */ + length?: string; + /** Codec used to compress Journal Fragments. */ + compressionCodec?: ProtocolCompressionCodec; + /** + * Storage backend base path for this Journal's Fragments. Must be in URL + * form, with the choice of backend defined by the scheme. The full path of + * a Journal's Fragment is derived by joining the store path with the + * Fragment's ContentPath. Eg, given a fragment_store of + * "s3://My-AWS-bucket/a/prefix" and a JournalSpec of name "my/journal", + * a complete Fragment path might be: + * "s3://My-AWS-bucket/a/prefix/my/journal/000123-000456-789abcdef.gzip + * Multiple stores may be specified, in which case the Journal's Fragments + * are the union of all Fragments present across all stores, and new + * Fragments always persist to the first specified store. This can be + * helpful in performing incremental migrations, where new Journal content + * is written to the new store, while content in the old store remains + * available (and, depending on fragment_retention or recovery log pruning, + * may eventually be removed). + * + * If no stores are specified, the Journal is still use-able but will + * not persist Fragments to any a backing fragment store. This allows for + * real-time streaming use cases where reads of historical data are not + * needed. + */ + stores?: string[]; + /** + * Interval of time between refreshes of remote Fragment listings from + * configured fragment_stores. + */ + refreshInterval?: string; + /** + * Retention duration for historical Fragments of this Journal within the + * Fragment stores. If less than or equal to zero, Fragments are retained + * indefinitely. + */ + retention?: string; + /** + * Flush interval defines a uniform UTC time segment which, when passed, + * will prompt brokers to close and persist a fragment presently being + * written. + * + * Flush interval may be helpful in integrating the journal with a regularly + * scheduled batch work-flow which processes new files from the fragment + * store and has no particular awareness of Gazette. For example, setting + * flush_interval to 3600s will cause brokers to persist their present + * fragment on the hour, every hour, even if it has not yet reached its + * target length. A batch work-flow running at 5 minutes past the hour is + * then reasonably assured of seeing all events from the past hour. + * See also "gazctl journals fragments --help" for more discussion. + */ + flushInterval?: string; + /** + * Path postfix template is a Go template which evaluates to a partial + * path under which fragments are persisted to the store. A complete + * fragment path is constructed by appending path components from the + * fragment store, then the journal name, and then the postfix template. + * Path post-fixes can help in maintaining Hive compatible partitioning + * over fragment creation time. The fields ".Spool" and ".JournalSpec" + * are available for introspection in the template. For example, + * to partition on the UTC date and hour of creation, use: + * date={{ .Spool.FirstAppendTime.Format "2006-01-02" }}/hour={{ + * .Spool.FirstAppendTime.Format "15" }} + * + * Which will produce a path postfix like "date=2019-11-19/hour=22". + */ + pathPostfixTemplate?: string; } /** - * Label defines a key & value pair which can be attached to entities like - * JournalSpecs and BrokerSpecs. Labels may be used to provide identifying - * attributes which do not directly imply semantics to the core system, but - * are meaningful to users or for higher-level Gazette tools. - */ +* Label defines a key & value pair which can be attached to entities like +JournalSpecs and BrokerSpecs. Labels may be used to provide identifying +attributes which do not directly imply semantics to the core system, but +are meaningful to users or for higher-level Gazette tools. +*/ export interface ProtocolLabel { - name?: string; - value?: string; + name?: string; + value?: string; } -/** LabelSelector defines a filter over LabelSets. */ +/** + * LabelSelector defines a filter over LabelSets. + */ export interface ProtocolLabelSelector { - /** - * Include is Labels which must be matched for a LabelSet to be selected. If - * empty, all Labels are included. An include Label with empty ("") value is - * matched by a Label of the same name having any value. - */ - include?: ProtocolLabelSet; - /** - * Exclude is Labels which cannot be matched for a LabelSet to be selected. If - * empty, no Labels are excluded. An exclude Label with empty ("") value - * excludes a Label of the same name having any value. - */ - exclude?: ProtocolLabelSet; + /** + * Include is Labels which must be matched for a LabelSet to be selected. If + * empty, all Labels are included. An include Label with empty ("") value is + * matched by a Label of the same name having any value. + */ + include?: ProtocolLabelSet; + /** + * Exclude is Labels which cannot be matched for a LabelSet to be selected. If + * empty, no Labels are excluded. An exclude Label with empty ("") value + * excludes a Label of the same name having any value. + */ + exclude?: ProtocolLabelSet; } -/** LabelSet is a collection of labels and their values. */ +/** + * LabelSet is a collection of labels and their values. + */ export interface ProtocolLabelSet { - /** Labels of the set. Instances must be unique and sorted over (Name, Value). */ - labels?: ProtocolLabel[]; + /** Labels of the set. Instances must be unique and sorted over (Name, Value). */ + labels?: ProtocolLabel[]; } -/** ListRequest is the unary request message of the broker List RPC. */ +/** + * ListRequest is the unary request message of the broker List RPC. + */ export interface ProtocolListRequest { - /** - * Selector optionally refines the set of journals which will be enumerated. - * If zero-valued, all journals are returned. Otherwise, only JournalSpecs - * matching the LabelSelector will be returned. Two meta-labels "name" and - * "prefix" are additionally supported by the selector, where: - * * name=examples/a-name will match a JournalSpec with Name - * "examples/a-name" - * * prefix=examples/ will match any JournalSpec having prefix "examples/". - * The prefix Label value must end in '/'. - */ - selector?: ProtocolLabelSelector; + /** + * Selector optionally refines the set of journals which will be enumerated. + * If zero-valued, all journals are returned. Otherwise, only JournalSpecs + * matching the LabelSelector will be returned. Two meta-labels "name" and + * "prefix" are additionally supported by the selector, where: + * * name=examples/a-name will match a JournalSpec with Name + * "examples/a-name" + * * prefix=examples/ will match any JournalSpec having prefix "examples/". + * The prefix Label value must end in '/'. + */ + selector?: ProtocolLabelSelector; } -/** ListResponse is the unary response message of the broker List RPC. */ +/** + * ListResponse is the unary response message of the broker List RPC. + */ export interface ProtocolListResponse { - /** Status of the List RPC. */ - status?: ProtocolStatus; - /** Header of the response. */ - header?: ProtocolHeader; - journals?: ListResponseJournal[]; + /** Status of the List RPC. */ + status?: ProtocolStatus; + /** Header of the response. */ + header?: ProtocolHeader; + journals?: ListResponseJournal[]; } -/** ReadRequest is the unary request message of the broker Read RPC. */ +/** + * ReadRequest is the unary request message of the broker Read RPC. + */ export interface ProtocolReadRequest { - /** Header is attached by a proxying broker peer. */ - header?: ProtocolHeader; - /** Journal to be read. */ - journal?: string; - /** - * Desired offset to begin reading from. Value -1 has special handling, where - * the read is performed from the current write head. All other positive - * values specify a desired exact byte offset to read from. If the offset is - * not available (eg, because it represents a portion of Journal which has - * been permanently deleted), the broker will return the next available - * offset. Callers should therefore always inspect the ReadResponse offset. - * @format int64 - */ - offset?: string; - /** - * Whether the operation should block until content becomes available. - * OFFSET_NOT_YET_AVAILABLE is returned if a non-blocking read has no ready - * content. - */ - block?: boolean; - /** - * If do_not_proxy is true, the broker will not proxy the read to another - * broker, or open and proxy a remote Fragment on the client's behalf. - */ - doNotProxy?: boolean; - /** - * If metadata_only is true, the broker will respond with Journal and - * Fragment metadata but not content. - */ - metadataOnly?: boolean; - /** - * Offset to read through. If zero, then the read end offset is unconstrained. - * @format int64 - */ - endOffset?: string; + /** Header is attached by a proxying broker peer. */ + header?: ProtocolHeader; + /** Journal to be read. */ + journal?: string; + /** + * Desired offset to begin reading from. Value -1 has special handling, where + * the read is performed from the current write head. All other positive + * values specify a desired exact byte offset to read from. If the offset is + * not available (eg, because it represents a portion of Journal which has + * been permanently deleted), the broker will return the next available + * offset. Callers should therefore always inspect the ReadResponse offset. + * @format int64 + */ + offset?: string; + /** + * Whether the operation should block until content becomes available. + * OFFSET_NOT_YET_AVAILABLE is returned if a non-blocking read has no ready + * content. + */ + block?: boolean; + /** + * If do_not_proxy is true, the broker will not proxy the read to another + * broker, or open and proxy a remote Fragment on the client's behalf. + */ + doNotProxy?: boolean; + /** + * If metadata_only is true, the broker will respond with Journal and + * Fragment metadata but not content. + */ + metadataOnly?: boolean; + /** + * Offset to read through. If zero, then the read end offset is unconstrained. + * @format int64 + */ + endOffset?: string; + /** + * BeginModTime is an optional inclusive lower bound on the modification + * timestamps of fragments read from the backing store, represented as + * seconds since the epoch. The request Offset will be advanced as-needed + * to skip persisted Fragments having a modication time before the bound. + * @format int64 + */ + beginModTime?: string; } /** - * ReadResponse is the streamed response message of the broker Read RPC. - * Responses messages are of two types: - * * "Metadata" messages, which conveys the journal Fragment addressed by the - * request which is ready to be read. - * * "Chunk" messages, which carry associated journal Fragment content bytes. - * - * A metadata message specifying a Fragment always precedes all "chunks" of the - * Fragment's content. Response streams may be very long lived, having many - * metadata and accompanying chunk messages. The reader may also block for long - * periods of time awaiting the next metadata message (eg, if the next offset - * hasn't yet committed). However once a metadata message is read, the reader - * is assured that its associated chunk messages are immediately forthcoming. - */ +* ReadResponse is the streamed response message of the broker Read RPC. +Responses messages are of two types: +* * "Metadata" messages, which conveys the journal Fragment addressed by the + request which is ready to be read. +* "Chunk" messages, which carry associated journal Fragment content bytes. + +A metadata message specifying a Fragment always precedes all "chunks" of the +Fragment's content. Response streams may be very long lived, having many +metadata and accompanying chunk messages. The reader may also block for long +periods of time awaiting the next metadata message (eg, if the next offset +hasn't yet committed). However once a metadata message is read, the reader +is assured that its associated chunk messages are immediately forthcoming. +*/ export interface ProtocolReadResponse { - /** Status of the Read RPC. */ - status?: ProtocolStatus; - /** - * Header of the response. Accompanies the first ReadResponse of the response - * stream. - */ - header?: ProtocolHeader; - /** - * The effective offset of the read. See ReadRequest offset. - * @format int64 - */ - offset?: string; - /** - * The offset to next be written, by the next append transaction served by - * broker. In other words, the last offset through which content is - * available to be read from the Journal. This is a metadata field and will - * not be returned with a content response. - * @format int64 - */ - writeHead?: string; - /** - * Fragment to which the offset was mapped. This is a metadata field and will - * not be returned with a content response. - */ - fragment?: ProtocolFragment; - /** If Fragment is remote, a URL from which it may be directly read. */ - fragmentUrl?: string; - /** - * Content chunks of the read. - * @format byte - */ - content?: string; + /** Status of the Read RPC. */ + status?: ProtocolStatus; + /** + * Header of the response. Accompanies the first ReadResponse of the response + * stream. + */ + header?: ProtocolHeader; + /** + * The effective offset of the read. See ReadRequest offset. + * @format int64 + */ + offset?: string; + /** + * The offset to next be written, by the next append transaction served by + * broker. In other words, the last offset through which content is + * available to be read from the Journal. This is a metadata field and will + * not be returned with a content response. + * @format int64 + */ + writeHead?: string; + /** + * Fragment to which the offset was mapped. This is a metadata field and will + * not be returned with a content response. + */ + fragment?: ProtocolFragment; + /** If Fragment is remote, a URL from which it may be directly read. */ + fragmentUrl?: string; + /** + * Content chunks of the read. + * @format byte + */ + content?: string; } /** - * ReplicateResponse is the streamed response message of the broker's internal - * Replicate RPC. Each message is a 1:1 response to a previously read "proposal" - * ReplicateRequest with |acknowledge| set. - */ +* ReplicateResponse is the streamed response message of the broker's internal +Replicate RPC. Each message is a 1:1 response to a previously read "proposal" +ReplicateRequest with |acknowledge| set. +*/ export interface ProtocolReplicateResponse { - /** Status of the Replicate RPC. */ - status?: ProtocolStatus; - /** - * Header of the response. Accompanies the first ReplicateResponse of the - * response stream. - */ - header?: ProtocolHeader; - /** - * If status is PROPOSAL_MISMATCH, then |fragment| is the replica's current - * journal Fragment, and either it or |registers| will differ from the - * primary's proposal. - */ - fragment?: ProtocolFragment; - /** - * If status is PROPOSAL_MISMATCH, then |registers| are the replica's current - * journal registers. - */ - registers?: ProtocolLabelSet; + /** Status of the Replicate RPC. */ + status?: ProtocolStatus; + /** + * Header of the response. Accompanies the first ReplicateResponse of the + * response stream. + */ + header?: ProtocolHeader; + /** + * If status is PROPOSAL_MISMATCH, then |fragment| is the replica's current + * journal Fragment, and either it or |registers| will differ from the + * primary's proposal. + */ + fragment?: ProtocolFragment; + /** + * If status is PROPOSAL_MISMATCH, then |registers| are the replica's current + * journal registers. + */ + registers?: ProtocolLabelSet; } -/** Route captures the current topology of an item and the processes serving it. */ +/** + * Route captures the current topology of an item and the processes serving it. + */ export interface ProtocolRoute { - /** Members of the Route, ordered on ascending ProcessSpec.ID (zone, suffix). */ - members?: ProcessSpecID[]; - /** - * Index of the ProcessSpec serving as primary within |members|, - * or -1 of no member is currently primary. - * @format int32 - */ - primary?: number; - /** - * Endpoints of each Route member. If not empty, |endpoints| has the same - * length and order as |members|, and captures the endpoint of each one. - */ - endpoints?: string[]; + /** Members of the Route, ordered on ascending ProcessSpec.ID (zone, suffix). */ + members?: ProcessSpecID[]; + /** + * Index of the ProcessSpec serving as primary within |members|, + * or -1 of no member is currently primary. + * @format int32 + */ + primary?: number; + /** + * Endpoints of each Route member. If not empty, |endpoints| has the same + * length and order as |members|, and captures the endpoint of each one. + */ + endpoints?: string[]; } -/** SHA1Sum is a 160-bit SHA1 digest. */ +/** + * SHA1Sum is a 160-bit SHA1 digest. + */ export interface ProtocolSHA1Sum { - /** @format uint64 */ - part1?: string; - /** @format uint64 */ - part2?: string; - /** @format int64 */ - part3?: number; + /** @format uint64 */ + part1?: string; + /** @format uint64 */ + part2?: string; + /** @format int64 */ + part3?: number; } /** - * Status is a response status code, used universally across Gazette RPC APIs. - * - * - JOURNAL_NOT_FOUND: The named journal does not exist. - * - NO_JOURNAL_PRIMARY_BROKER: There is no current primary broker for the journal. This is a temporary - * condition which should quickly resolve, assuming sufficient broker - * capacity. - * - NOT_JOURNAL_PRIMARY_BROKER: The present broker is not the assigned primary broker for the journal. - * - NOT_JOURNAL_BROKER: The present broker is not an assigned broker for the journal. - * - INSUFFICIENT_JOURNAL_BROKERS: There are an insufficient number of assigned brokers for the journal - * to meet its required replication. - * - OFFSET_NOT_YET_AVAILABLE: The requested offset is not yet available. This indicates either that the - * offset has not yet been written, or that the broker is not yet aware of a - * written fragment covering the offset. Returned only by non-blocking reads. - * - WRONG_ROUTE: The peer disagrees with the Route accompanying a ReplicateRequest. - * - PROPOSAL_MISMATCH: The peer disagrees with the proposal accompanying a ReplicateRequest. - * - ETCD_TRANSACTION_FAILED: The Etcd transaction failed. Returned by Update RPC when an - * expect_mod_revision of the UpdateRequest differs from the current - * ModRevision of the JournalSpec within the store. - * - NOT_ALLOWED: A disallowed journal access was attempted (eg, a write where the - * journal disables writes, or read where journals disable reads). - * - WRONG_APPEND_OFFSET: The Append is refused because its requested offset is not equal - * to the furthest written offset of the journal. - * - INDEX_HAS_GREATER_OFFSET: The Append is refused because the replication pipeline tracks a smaller - * journal offset than that of the remote fragment index. This indicates - * that journal replication consistency has been lost in the past, due to - * too many broker or Etcd failures. - * - REGISTER_MISMATCH: The Append is refused because a registers selector was provided with the - * request, but it was not matched by current register values of the journal. - * @default "OK" - */ +* Status is a response status code, used universally across Gazette RPC APIs. + + - JOURNAL_NOT_FOUND: The named journal does not exist. + - NO_JOURNAL_PRIMARY_BROKER: There is no current primary broker for the journal. This is a temporary +condition which should quickly resolve, assuming sufficient broker +capacity. + - NOT_JOURNAL_PRIMARY_BROKER: The present broker is not the assigned primary broker for the journal. + - NOT_JOURNAL_BROKER: The present broker is not an assigned broker for the journal. + - INSUFFICIENT_JOURNAL_BROKERS: There are an insufficient number of assigned brokers for the journal +to meet its required replication. + - OFFSET_NOT_YET_AVAILABLE: The requested offset is not yet available. This indicates either that the +offset has not yet been written, or that the broker is not yet aware of a +written fragment covering the offset. Returned only by non-blocking reads. + - WRONG_ROUTE: The peer disagrees with the Route accompanying a ReplicateRequest. + - PROPOSAL_MISMATCH: The peer disagrees with the proposal accompanying a ReplicateRequest. + - ETCD_TRANSACTION_FAILED: The Etcd transaction failed. Returned by Update RPC when an +expect_mod_revision of the UpdateRequest differs from the current +ModRevision of the JournalSpec within the store. + - NOT_ALLOWED: A disallowed journal access was attempted (eg, a write where the +journal disables writes, or read where journals disable reads). + - WRONG_APPEND_OFFSET: The Append is refused because its requested offset is not equal +to the furthest written offset of the journal. + - INDEX_HAS_GREATER_OFFSET: The Append is refused because the replication pipeline tracks a smaller +journal offset than that of the remote fragment index. This indicates +that journal replication consistency has been lost in the past, due to +too many broker or Etcd failures. + - REGISTER_MISMATCH: The Append is refused because a registers selector was provided with the +request, but it was not matched by current register values of the journal. +*/ export type ProtocolStatus = - | "OK" - | "JOURNAL_NOT_FOUND" - | "NO_JOURNAL_PRIMARY_BROKER" - | "NOT_JOURNAL_PRIMARY_BROKER" - | "NOT_JOURNAL_BROKER" - | "INSUFFICIENT_JOURNAL_BROKERS" - | "OFFSET_NOT_YET_AVAILABLE" - | "WRONG_ROUTE" - | "PROPOSAL_MISMATCH" - | "ETCD_TRANSACTION_FAILED" - | "NOT_ALLOWED" - | "WRONG_APPEND_OFFSET" - | "INDEX_HAS_GREATER_OFFSET" - | "REGISTER_MISMATCH"; + | 'OK' + | 'JOURNAL_NOT_FOUND' + | 'NO_JOURNAL_PRIMARY_BROKER' + | 'NOT_JOURNAL_PRIMARY_BROKER' + | 'NOT_JOURNAL_BROKER' + | 'INSUFFICIENT_JOURNAL_BROKERS' + | 'OFFSET_NOT_YET_AVAILABLE' + | 'WRONG_ROUTE' + | 'PROPOSAL_MISMATCH' + | 'ETCD_TRANSACTION_FAILED' + | 'NOT_ALLOWED' + | 'WRONG_APPEND_OFFSET' + | 'INDEX_HAS_GREATER_OFFSET' + | 'REGISTER_MISMATCH'; export interface RuntimeError { - error?: string; - /** @format int32 */ - code?: number; - message?: string; - details?: ProtobufAny[]; + error?: string; + /** @format int32 */ + code?: number; + message?: string; + details?: ProtobufAny[]; } export interface RuntimeStreamError { - /** @format int32 */ - grpcCode?: number; - /** @format int32 */ - httpCode?: number; - message?: string; - httpStatus?: string; - details?: ProtobufAny[]; + /** @format int32 */ + grpcCode?: number; + /** @format int32 */ + httpCode?: number; + message?: string; + httpStatus?: string; + details?: ProtobufAny[]; } diff --git a/client/dist/src/gen/consumer/protocol/consumer.ts b/client/dist/src/gen/consumer/protocol/consumer.ts index e39de6f..e8c797e 100644 --- a/client/dist/src/gen/consumer/protocol/consumer.ts +++ b/client/dist/src/gen/consumer/protocol/consumer.ts @@ -10,663 +10,669 @@ */ export interface GetHintsResponseResponseHints { - /** If the hints value does not exist Hints will be nil. */ - hints?: RecoverylogFSMHints; + /** If the hints value does not exist Hints will be nil. */ + hints?: RecoverylogFSMHints; } /** - * Etcd represents the effective Etcd MVCC state under which a Gazette broker - * is operating in its processing of requests and responses. Its inclusion - * allows brokers to reason about relative "happened before" Revision ordering - * of apparent routing conflicts in proxied or replicated requests, as well - * as enabling sanity checks over equality of Etcd ClusterId (and precluding, - * for example, split-brain scenarios where different brokers are backed by - * different Etcd clusters). Etcd is kept in sync with - * etcdserverpb.ResponseHeader. - */ +* Etcd represents the effective Etcd MVCC state under which a Gazette broker +is operating in its processing of requests and responses. Its inclusion +allows brokers to reason about relative "happened before" Revision ordering +of apparent routing conflicts in proxied or replicated requests, as well +as enabling sanity checks over equality of Etcd ClusterId (and precluding, +for example, split-brain scenarios where different brokers are backed by +different Etcd clusters). Etcd is kept in sync with +etcdserverpb.ResponseHeader. +*/ export interface HeaderEtcd { - /** - * cluster_id is the ID of the cluster. - * @format uint64 - */ - clusterId?: string; - /** - * member_id is the ID of the member. - * @format uint64 - */ - memberId?: string; - /** - * revision is the Etcd key-value store revision when the request was - * applied. - * @format int64 - */ - revision?: string; - /** - * raft_term is the raft term when the request was applied. - * @format uint64 - */ - raftTerm?: string; + /** + * cluster_id is the ID of the cluster. + * @format uint64 + */ + clusterId?: string; + /** + * member_id is the ID of the member. + * @format uint64 + */ + memberId?: string; + /** + * revision is the Etcd key-value store revision when the request was + * applied. + * @format int64 + */ + revision?: string; + /** + * raft_term is the raft term when the request was applied. + * @format uint64 + */ + raftTerm?: string; } -/** Shards of the response. */ +/** + * Shards of the response. + */ export interface ListResponseShard { - /** - * ShardSpec describes a shard and its configuration, and is the long-lived unit - * of work and scaling for a consumer application. Each shard is allocated to a - * one "primary" at-a-time selected from the current processes of a consumer - * application, and is re-assigned on process fault or exit. - * - * ShardSpecs describe all configuration of the shard and its processing, - * including journals to consume, configuration for processing transactions, its - * recovery log, hot standbys, etc. ShardSpecs may be further extended with - * domain-specific labels & values to further define application behavior. - * ShardSpec is-a allocator.ItemValue. - */ - spec?: ConsumerShardSpec; - /** - * Current ModRevision of the ShardSpec. - * @format int64 - */ - modRevision?: string; - /** Route of the shard, including endpoints. */ - route?: ProtocolRoute; - /** Status of each replica. Cardinality and ordering matches |route|. */ - status?: ConsumerReplicaStatus[]; + /** + * ShardSpec describes a shard and its configuration, and is the long-lived unit + * of work and scaling for a consumer application. Each shard is allocated to a + * one "primary" at-a-time selected from the current processes of a consumer + * application, and is re-assigned on process fault or exit. + * + * ShardSpecs describe all configuration of the shard and its processing, + * including journals to consume, configuration for processing transactions, its + * recovery log, hot standbys, etc. ShardSpecs may be further extended with + * domain-specific labels & values to further define application behavior. + * ShardSpec is-a allocator.ItemValue. + */ + spec?: ConsumerShardSpec; + /** + * Current ModRevision of the ShardSpec. + * @format int64 + */ + modRevision?: string; + /** Route of the shard, including endpoints. */ + route?: ProtocolRoute; + /** Status of each replica. Cardinality and ordering matches |route|. */ + status?: ConsumerReplicaStatus[]; } -/** ID composes a zone and a suffix to uniquely identify a ProcessSpec. */ +/** + * ID composes a zone and a suffix to uniquely identify a ProcessSpec. + */ export interface ProcessSpecID { - /** - * "Zone" in which the process is running. Zones may be AWS, Azure, or - * Google Cloud Platform zone identifiers, or rack locations within a colo, - * or given some other custom meaning. Gazette will replicate across - * multiple zones, and seeks to minimize traffic which must cross zones (for - * example, by proxying reads to a broker in the current zone). - */ - zone?: string; - /** - * Unique suffix of the process within |zone|. It is permissible for a - * suffix value to repeat across zones, but never within zones. In practice, - * it's recommended to use a FQDN, Kubernetes Pod name, or comparable unique - * and self-describing value as the ID suffix. - */ - suffix?: string; + /** + * "Zone" in which the process is running. Zones may be AWS, Azure, or + * Google Cloud Platform zone identifiers, or rack locations within a colo, + * or given some other custom meaning. Gazette will replicate across + * multiple zones, and seeks to minimize traffic which must cross zones (for + * example, by proxying reads to a broker in the current zone). + */ + zone?: string; + /** + * Unique suffix of the process within |zone|. It is permissible for a + * suffix value to repeat across zones, but never within zones. In practice, + * it's recommended to use a FQDN, Kubernetes Pod name, or comparable unique + * and self-describing value as the ID suffix. + */ + suffix?: string; } /** - * - BACKFILL: The replica is actively playing the historical recovery log. - * - STANDBY: The replica has finished playing the historical recovery log and is - * live-tailing it to locally mirror recorded operations as they are - * produced. It can take over as primary at any time. - * - * Shards not having recovery logs immediately transition to STANDBY. - * - PRIMARY: The replica is actively serving as primary. - * - FAILED: The replica has encountered an unrecoverable error. - * @default "IDLE" - */ -export type ReplicaStatusCode = "IDLE" | "BACKFILL" | "STANDBY" | "PRIMARY" | "FAILED"; +* - BACKFILL: The replica is actively playing the historical recovery log. + - STANDBY: The replica has finished playing the historical recovery log and is +live-tailing it to locally mirror recorded operations as they are +produced. It can take over as primary at any time. + +Shards not having recovery logs immediately transition to STANDBY. + - PRIMARY: The replica is actively serving as primary. + - FAILED: The replica has encountered an unrecoverable error. +*/ +export type ReplicaStatusCode = 'IDLE' | 'BACKFILL' | 'STANDBY' | 'PRIMARY' | 'FAILED'; /** - * Change defines an insertion, update, or deletion to be applied to the set - * of ShardSpecs. Exactly one of |upsert| or |delete| must be set. - */ +* Change defines an insertion, update, or deletion to be applied to the set +of ShardSpecs. Exactly one of |upsert| or |delete| must be set. +*/ export interface ConsumerApplyRequestChange { - /** - * Expected ModRevision of the current ShardSpec. If the shard is being - * created, expect_mod_revision is zero. - * @format int64 - */ - expectModRevision?: string; - /** - * ShardSpec to be updated (if expect_mod_revision > 0) or created - * (if expect_mod_revision == 0). - */ - upsert?: ConsumerShardSpec; - /** Shard to be deleted. expect_mod_revision must not be zero. */ - delete?: string; + /** + * Expected ModRevision of the current ShardSpec. If the shard is being + * created, expect_mod_revision is zero. + * @format int64 + */ + expectModRevision?: string; + /** + * ShardSpec to be updated (if expect_mod_revision > 0) or created + * (if expect_mod_revision == 0). + */ + upsert?: ConsumerShardSpec; + /** Shard to be deleted. expect_mod_revision must not be zero. */ + delete?: string; } export interface ConsumerApplyResponse { - /** Status of the Apply RPC. */ - status?: ConsumerStatus; - /** Header of the response. */ - header?: ProtocolHeader; - /** - * Optional extension of the ApplyResponse. - * @format byte - */ - extension?: string; + /** Status of the Apply RPC. */ + status?: ConsumerStatus; + /** Header of the response. */ + header?: ProtocolHeader; + /** + * Optional extension of the ApplyResponse. + * @format byte + */ + extension?: string; } export interface ConsumerGetHintsResponse { - /** Status of the Hints RPC. */ - status?: ConsumerStatus; - /** Header of the response. */ - header?: ProtocolHeader; - /** Primary hints for the shard. */ - primaryHints?: GetHintsResponseResponseHints; - /** - * List of backup hints for a shard. The most recent recovery log hints will - * be first, any subsequent hints are for historical backup. If there is no - * value for a hint key the value corresponding hints will be nil. - */ - backupHints?: GetHintsResponseResponseHints[]; - /** - * Optional extension of the GetHintsResponse. - * @format byte - */ - extension?: string; + /** Status of the Hints RPC. */ + status?: ConsumerStatus; + /** Header of the response. */ + header?: ProtocolHeader; + /** Primary hints for the shard. */ + primaryHints?: GetHintsResponseResponseHints; + /** + * List of backup hints for a shard. The most recent recovery log hints will + * be first, any subsequent hints are for historical backup. If there is no + * value for a hint key the value corresponding hints will be nil. + */ + backupHints?: GetHintsResponseResponseHints[]; + /** + * Optional extension of the GetHintsResponse. + * @format byte + */ + extension?: string; } export interface ConsumerListRequest { - /** - * Selector optionally refines the set of shards which will be enumerated. - * If zero-valued, all shards are returned. Otherwise, only ShardSpecs - * matching the LabelSelector will be returned. One meta-label "id" is - * additionally supported by the selector, where "id=example-shard-ID" - * will match a ShardSpec with ID "example-shard-ID". - */ - selector?: ProtocolLabelSelector; - /** - * Optional extension of the ListRequest. - * @format byte - */ - extension?: string; + /** + * Selector optionally refines the set of shards which will be enumerated. + * If zero-valued, all shards are returned. Otherwise, only ShardSpecs + * matching the LabelSelector will be returned. One meta-label "id" is + * additionally supported by the selector, where "id=example-shard-ID" + * will match a ShardSpec with ID "example-shard-ID". + */ + selector?: ProtocolLabelSelector; + /** + * Optional extension of the ListRequest. + * @format byte + */ + extension?: string; } export interface ConsumerListResponse { - /** Status of the List RPC. */ - status?: ConsumerStatus; - /** Header of the response. */ - header?: ProtocolHeader; - shards?: ListResponseShard[]; - /** - * Optional extension of the ListResponse. - * @format byte - */ - extension?: string; + /** Status of the List RPC. */ + status?: ConsumerStatus; + /** Header of the response. */ + header?: ProtocolHeader; + shards?: ListResponseShard[]; + /** + * Optional extension of the ListResponse. + * @format byte + */ + extension?: string; } /** - * ReplicaStatus is the status of a ShardSpec assigned to a ConsumerSpec. - * It serves as an allocator AssignmentValue. ReplicaStatus is reduced by taking - * the maximum enum value among statuses. Eg, if a primary is PRIMARY, one - * replica is BACKFILL and the other STANDBY, then the status is PRIMARY. If one - * of the replicas transitioned to FAILED, than the status is FAILED. This - * reduction behavior is used to summarize status across all replicas. - */ +* ReplicaStatus is the status of a ShardSpec assigned to a ConsumerSpec. +It serves as an allocator AssignmentValue. ReplicaStatus is reduced by taking +the maximum enum value among statuses. Eg, if a primary is PRIMARY, one +replica is BACKFILL and the other STANDBY, then the status is PRIMARY. If one +of the replicas transitioned to FAILED, than the status is FAILED. This +reduction behavior is used to summarize status across all replicas. +*/ export interface ConsumerReplicaStatus { - /** - * - BACKFILL: The replica is actively playing the historical recovery log. - * - STANDBY: The replica has finished playing the historical recovery log and is - * live-tailing it to locally mirror recorded operations as they are - * produced. It can take over as primary at any time. - * - * Shards not having recovery logs immediately transition to STANDBY. - * - PRIMARY: The replica is actively serving as primary. - * - FAILED: The replica has encountered an unrecoverable error. - */ - code?: ReplicaStatusCode; - /** Errors encountered during replica processing. Set iff |code| is FAILED. */ - errors?: string[]; + /** + * - BACKFILL: The replica is actively playing the historical recovery log. + * - STANDBY: The replica has finished playing the historical recovery log and is + * live-tailing it to locally mirror recorded operations as they are + * produced. It can take over as primary at any time. + * + * Shards not having recovery logs immediately transition to STANDBY. + * - PRIMARY: The replica is actively serving as primary. + * - FAILED: The replica has encountered an unrecoverable error. + */ + code?: ReplicaStatusCode; + /** Errors encountered during replica processing. Set iff |code| is FAILED. */ + errors?: string[]; } /** - * ShardSpec describes a shard and its configuration, and is the long-lived unit - * of work and scaling for a consumer application. Each shard is allocated to a - * one "primary" at-a-time selected from the current processes of a consumer - * application, and is re-assigned on process fault or exit. - * - * ShardSpecs describe all configuration of the shard and its processing, - * including journals to consume, configuration for processing transactions, its - * recovery log, hot standbys, etc. ShardSpecs may be further extended with - * domain-specific labels & values to further define application behavior. - * ShardSpec is-a allocator.ItemValue. - */ +* ShardSpec describes a shard and its configuration, and is the long-lived unit +of work and scaling for a consumer application. Each shard is allocated to a +one "primary" at-a-time selected from the current processes of a consumer +application, and is re-assigned on process fault or exit. + +ShardSpecs describe all configuration of the shard and its processing, +including journals to consume, configuration for processing transactions, its +recovery log, hot standbys, etc. ShardSpecs may be further extended with +domain-specific labels & values to further define application behavior. +ShardSpec is-a allocator.ItemValue. +*/ export interface ConsumerShardSpec { - /** ID of the shard. */ - id?: string; - /** Sources of the shard, uniquely ordered on Source journal. */ - sources?: ConsumerShardSpecSource[]; - /** - * Prefix of the Journal into which the shard's recovery log will be recorded. - * The complete Journal name is built as "{recovery_log_prefix}/{shard_id}". - * If empty, the shard does not use a recovery log. - */ - recoveryLogPrefix?: string; - /** - * Prefix of Etcd keys into which recovery log FSMHints are written to and - * read from. FSMHints allow readers of the recovery log to efficiently - * determine the minimum fragments of log which must be read to fully recover - * local store state. The complete hint key written by the shard primary is: - * "{hint_prefix}/{shard_id}.primary" - * - * The primary will regularly produce updated hints into this key, and - * players of the log will similarly utilize hints from this key. - * If |recovery_log_prefix| is set, |hint_prefix| must be also. - */ - hintPrefix?: string; - /** - * Backups of verified recovery log FSMHints, retained as a disaster-recovery - * mechanism. On completing playback, a player will write recovered hints to: - * "{hints_prefix}/{shard_id}.backup.0". - * - * It also move hints previously stored under - * "{hints_prefix/{shard_id}.backup.0" to - * "{hints_prefix/{shard_id}.backup.1", and so on, keeping at most - * |hint_backups| distinct sets of FSMHints. - * - * In the case of disaster or data-loss, these copied hints can be an - * important fallback for recovering a consistent albeit older version of the - * shard's store, with each relying on only progressively older portions of - * the recovery log. - * - * When pruning the recovery log, log fragments which are older than (and no - * longer required by) the *oldest* backup are discarded, ensuring that - * all hints remain valid for playback. - * @format int32 - */ - hintBackups?: number; - /** - * Max duration of shard transactions. This duration upper-bounds the amount - * of time during which a transaction may process messages before it must - * flush and commit. It may run for less time if an input message stall occurs - * (eg, no decoded journal message is ready without blocking). A typical value - * would be `1s`: applications which perform extensive aggregation over - * message streams exhibiting locality of "hot" keys may benefit from larger - * values. - */ - maxTxnDuration?: string; - /** - * Min duration of shard transactions. This duration lower-bounds the amount - * of time during which a transaction must process messages before it may - * flush and commit. It may run for more time if additional messages are - * available (eg, decoded journal messages are ready without blocking). Note - * also that transactions are pipelined: a current transaction may process - * messages while a prior transaction's recovery log writes flush to Gazette, - * but it cannot prepare to commit until the prior transaction writes - * complete. In other words even if |min_txn_quantum| is zero, some degree of - * message batching is expected due to the network delay inherent in Gazette - * writes. A typical value of would be `0s`: applications which perform - * extensive aggregation may benefit from larger values. - */ - minTxnDuration?: string; - /** Disable processing of the shard. */ - disable?: boolean; - /** - * Hot standbys is the desired number of consumer processes which should be - * replicating the primary consumer's recovery log. Standbys are allocated in - * a separate availability zone of the current primary, and tail the live log - * to continuously mirror the primary's on-disk DB file structure. Should the - * primary experience failure, one of the hot standbys will be assigned to - * take over as the new shard primary, which is accomplished by simply opening - * its local copy of the recovered store files. - * - * Note that under regular operation, shard hand-off is zero downtime even if - * standbys are zero, as the current primary will not cede ownership until the - * replacement process declares itself ready. However, without standbys a - * process failure will leave the shard without an active primary while its - * replacement starts and completes playback of its recovery log. - * @format int64 - */ - hotStandbys?: number; - /** - * User-defined Labels of this ShardSpec. The label "id" is reserved and may - * not be used with a ShardSpec's labels. - */ - labels?: ProtocolLabelSet; - /** - * Disable waiting for acknowledgements of pending message(s). - * - * If a consumer transaction reads uncommitted messages, it will by default - * remain open (subject to the max duration) awaiting an acknowledgement of - * those messages, in the hope that that acknowledgement will be quickly - * forthcoming and, by remaining open, we can process all messages in this - * transaction. Effectively we're trading a small amount of increased local - * latency for a global reduction in end-to-end latency. - * - * This works well for acyclic message flows, but can introduce unnecessary - * stalls if there are message cycles between shards. In the simplest case, - * a transaction could block awaiting an ACK of a message that it itself - * produced -- an ACK which can't arrive until the transaction closes. - */ - disableWaitForAck?: boolean; - /** - * Size of the ring buffer used to sequence read-uncommitted messages - * into consumed, read-committed ones. The ring buffer is a performance - * optimization only: applications will replay portions of journals as - * needed when messages aren't available in the buffer. - * It can remain small if source journal transactions are small, - * but larger transactions will achieve better performance with a - * larger ring. - * If zero, a reasonable default (currently 8192) is used. - * @format int64 - */ - ringBufferSize?: number; - /** - * Size of the channel used to bridge message read and decode with - * sequencing and consumption. Larger values may reduce data stalls, - * particularly for larger transactions and/or bursty custom - * MessageProducer implementations. - * If zero, a reasonable default (currently 8192) is used. - * @format int64 - */ - readChannelSize?: number; + /** ID of the shard. */ + id?: string; + /** Sources of the shard, uniquely ordered on Source journal. */ + sources?: ConsumerShardSpecSource[]; + /** + * Prefix of the Journal into which the shard's recovery log will be recorded. + * The complete Journal name is built as "{recovery_log_prefix}/{shard_id}". + * If empty, the shard does not use a recovery log. + */ + recoveryLogPrefix?: string; + /** + * Prefix of Etcd keys into which recovery log FSMHints are written to and + * read from. FSMHints allow readers of the recovery log to efficiently + * determine the minimum fragments of log which must be read to fully recover + * local store state. The complete hint key written by the shard primary is: + * "{hint_prefix}/{shard_id}.primary" + * + * The primary will regularly produce updated hints into this key, and + * players of the log will similarly utilize hints from this key. + * If |recovery_log_prefix| is set, |hint_prefix| must be also. + */ + hintPrefix?: string; + /** + * Backups of verified recovery log FSMHints, retained as a disaster-recovery + * mechanism. On completing playback, a player will write recovered hints to: + * "{hints_prefix}/{shard_id}.backup.0". + * + * It also move hints previously stored under + * "{hints_prefix/{shard_id}.backup.0" to + * "{hints_prefix/{shard_id}.backup.1", and so on, keeping at most + * |hint_backups| distinct sets of FSMHints. + * In the case of disaster or data-loss, these copied hints can be an + * important fallback for recovering a consistent albeit older version of the + * shard's store, with each relying on only progressively older portions of + * the recovery log. + * When pruning the recovery log, log fragments which are older than (and no + * longer required by) the *oldest* backup are discarded, ensuring that + * all hints remain valid for playback. + * @format int32 + */ + hintBackups?: number; + /** + * Max duration of shard transactions. This duration upper-bounds the amount + * of time during which a transaction may process messages before it must + * flush and commit. It may run for less time if an input message stall occurs + * (eg, no decoded journal message is ready without blocking). A typical value + * would be `1s`: applications which perform extensive aggregation over + * message streams exhibiting locality of "hot" keys may benefit from larger + * values. + */ + maxTxnDuration?: string; + /** + * Min duration of shard transactions. This duration lower-bounds the amount + * of time during which a transaction must process messages before it may + * flush and commit. It may run for more time if additional messages are + * available (eg, decoded journal messages are ready without blocking). Note + * also that transactions are pipelined: a current transaction may process + * messages while a prior transaction's recovery log writes flush to Gazette, + * but it cannot prepare to commit until the prior transaction writes + * complete. In other words even if |min_txn_quantum| is zero, some degree of + * message batching is expected due to the network delay inherent in Gazette + * writes. A typical value of would be `0s`: applications which perform + * extensive aggregation may benefit from larger values. + */ + minTxnDuration?: string; + /** Disable processing of the shard. */ + disable?: boolean; + /** + * Hot standbys is the desired number of consumer processes which should be + * replicating the primary consumer's recovery log. Standbys are allocated in + * a separate availability zone of the current primary, and tail the live log + * to continuously mirror the primary's on-disk DB file structure. Should the + * primary experience failure, one of the hot standbys will be assigned to + * take over as the new shard primary, which is accomplished by simply opening + * its local copy of the recovered store files. + * + * Note that under regular operation, shard hand-off is zero downtime even if + * standbys are zero, as the current primary will not cede ownership until the + * replacement process declares itself ready. However, without standbys a + * process failure will leave the shard without an active primary while its + * replacement starts and completes playback of its recovery log. + * @format int64 + */ + hotStandbys?: number; + /** + * User-defined Labels of this ShardSpec. The label "id" is reserved and may + * not be used with a ShardSpec's labels. + */ + labels?: ProtocolLabelSet; + /** + * Disable waiting for acknowledgements of pending message(s). + * + * If a consumer transaction reads uncommitted messages, it will by default + * remain open (subject to the max duration) awaiting an acknowledgement of + * those messages, in the hope that that acknowledgement will be quickly + * forthcoming and, by remaining open, we can process all messages in this + * transaction. Effectively we're trading a small amount of increased local + * latency for a global reduction in end-to-end latency. + * This works well for acyclic message flows, but can introduce unnecessary + * stalls if there are message cycles between shards. In the simplest case, + * a transaction could block awaiting an ACK of a message that it itself + * produced -- an ACK which can't arrive until the transaction closes. + */ + disableWaitForAck?: boolean; + /** + * Size of the ring buffer used to sequence read-uncommitted messages + * into consumed, read-committed ones. The ring buffer is a performance + * optimization only: applications will replay portions of journals as + * needed when messages aren't available in the buffer. + * It can remain small if source journal transactions are small, + * but larger transactions will achieve better performance with a + * larger ring. + * If zero, a reasonable default (currently 8192) is used. + * @format int64 + */ + ringBufferSize?: number; + /** + * Size of the channel used to bridge message read and decode with + * sequencing and consumption. Larger values may reduce data stalls, + * particularly for larger transactions and/or bursty custom + * MessageProducer implementations. + * If zero, a reasonable default (currently 8192) is used. + * @format int64 + */ + readChannelSize?: number; } /** - * Sources define the set of journals which this shard consumes. At least one - * Source must be specified, and in many use cases only one will be needed. - * For use cases which can benefit, multiple sources may be specified to - * represent a "join" over messages of distinct journals. - * - * Note the effective mapping of messages to each of the joined journals - * should align (eg, joining a journal of customer updates with one of orders, - * where both are mapped on customer ID). This typically means the - * partitioning of the two event "topics" must be the same. - * - * Another powerful pattern is to shard on partitions of a high-volume event - * stream, and also have each shard join against all events of a low-volume - * stream. For example, a shard might ingest and index "viewed product" - * events, read a comparably low-volume "purchase" event stream, and on each - * purchase publish the bundle of its corresponding prior product views. - */ +* Sources define the set of journals which this shard consumes. At least one +Source must be specified, and in many use cases only one will be needed. +For use cases which can benefit, multiple sources may be specified to +represent a "join" over messages of distinct journals. + +Note the effective mapping of messages to each of the joined journals +should align (eg, joining a journal of customer updates with one of orders, +where both are mapped on customer ID). This typically means the +partitioning of the two event "topics" must be the same. + +Another powerful pattern is to shard on partitions of a high-volume event +stream, and also have each shard join against all events of a low-volume +stream. For example, a shard might ingest and index "viewed product" +events, read a comparably low-volume "purchase" event stream, and on each +purchase publish the bundle of its corresponding prior product views. +*/ export interface ConsumerShardSpecSource { - /** Journal which this shard is consuming. */ - journal?: string; - /** - * Minimum journal byte offset the shard should begin reading from. - * Typically this should be zero, as read offsets are check-pointed and - * restored from the shard's Store as it processes. |min_offset| can be - * useful for shard initialization, directing it to skip over historical - * portions of the journal not needed for the application's use case. - * @format int64 - */ - minOffset?: string; + /** Journal which this shard is consuming. */ + journal?: string; + /** + * Minimum journal byte offset the shard should begin reading from. + * Typically this should be zero, as read offsets are check-pointed and + * restored from the shard's Store as it processes. |min_offset| can be + * useful for shard initialization, directing it to skip over historical + * portions of the journal not needed for the application's use case. + * @format int64 + */ + minOffset?: string; } export interface ConsumerStatRequest { - /** Header may be attached by a proxying consumer peer. */ - header?: ProtocolHeader; - /** Shard to Stat. */ - shard?: string; - /** - * Journals and offsets which must be reflected in a completed consumer - * transaction before Stat returns, blocking if required. Offsets of journals - * not read by this shard are ignored. - */ - readThrough?: Record; - /** - * Optional extension of the StatRequest. - * @format byte - */ - extension?: string; + /** Header may be attached by a proxying consumer peer. */ + header?: ProtocolHeader; + /** Shard to Stat. */ + shard?: string; + /** + * Journals and offsets which must be reflected in a completed consumer + * transaction before Stat returns, blocking if required. Offsets of journals + * not read by this shard are ignored. + */ + readThrough?: Record; + /** + * Optional extension of the StatRequest. + * @format byte + */ + extension?: string; } export interface ConsumerStatResponse { - /** Status of the Stat RPC. */ - status?: ConsumerStatus; - /** Header of the response. */ - header?: ProtocolHeader; - /** - * Journals and offsets read through by the most recent completed consumer - * transaction. - */ - readThrough?: Record; - /** - * Journals and offsets this shard has published through, including - * acknowledgements, as-of the most recent completed consumer transaction. - * - * Formally, if an acknowledged message A results in this shard publishing - * messages B, and A falls within |read_through|, then all messages B & their - * acknowledgements fall within |publish_at|. - * - * The composition of |read_through| and |publish_at| allow CQRS applications - * to provide read-your-writes consistency, even if written events pass - * through multiple intermediate consumers and arbitrary transformations - * before arriving at the materialized view which is ultimately queried. - */ - publishAt?: Record; - /** - * Optional extension of the StatResponse. - * @format byte - */ - extension?: string; + /** Status of the Stat RPC. */ + status?: ConsumerStatus; + /** Header of the response. */ + header?: ProtocolHeader; + /** + * Journals and offsets read through by the most recent completed consumer + * transaction. + */ + readThrough?: Record; + /** + * Journals and offsets this shard has published through, including + * acknowledgements, as-of the most recent completed consumer transaction. + * + * Formally, if an acknowledged message A results in this shard publishing + * messages B, and A falls within |read_through|, then all messages B & their + * acknowledgements fall within |publish_at|. + * The composition of |read_through| and |publish_at| allow CQRS applications + * to provide read-your-writes consistency, even if written events pass + * through multiple intermediate consumers and arbitrary transformations + * before arriving at the materialized view which is ultimately queried. + */ + publishAt?: Record; + /** + * Optional extension of the StatResponse. + * @format byte + */ + extension?: string; } /** - * Status is a response status code, used across Gazette Consumer RPC APIs. - * - * - SHARD_NOT_FOUND: The named shard does not exist. - * - NO_SHARD_PRIMARY: There is no current primary consumer process for the shard. This is a - * temporary condition which should quickly resolve, assuming sufficient - * consumer capacity. - * - NOT_SHARD_PRIMARY: The present consumer process is not the assigned primary for the shard, - * and was not instructed to proxy the request. - * - ETCD_TRANSACTION_FAILED: The Etcd transaction failed. Returned by Update RPC when an - * expect_mod_revision of the UpdateRequest differs from the current - * ModRevision of the ShardSpec within the store. - * - SHARD_STOPPED: The current primary shard has stopped, either due to reassignment or - * processing failure, and will not make further progress toward the - * requested operation. - * For example, a Stat RPC will return SHARD_STOPPED if the StatRequest - * cannot be satisfied. - * @default "OK" - */ +* Status is a response status code, used across Gazette Consumer RPC APIs. + + - SHARD_NOT_FOUND: The named shard does not exist. + - NO_SHARD_PRIMARY: There is no current primary consumer process for the shard. This is a +temporary condition which should quickly resolve, assuming sufficient +consumer capacity. + - NOT_SHARD_PRIMARY: The present consumer process is not the assigned primary for the shard, +and was not instructed to proxy the request. + - ETCD_TRANSACTION_FAILED: The Etcd transaction failed. Returned by Update RPC when an +expect_mod_revision of the UpdateRequest differs from the current +ModRevision of the ShardSpec within the store. + - SHARD_STOPPED: The current primary shard has stopped, either due to reassignment or +processing failure, and will not make further progress toward the +requested operation. +For example, a Stat RPC will return SHARD_STOPPED if the StatRequest +cannot be satisfied. +*/ export type ConsumerStatus = - | "OK" - | "SHARD_NOT_FOUND" - | "NO_SHARD_PRIMARY" - | "NOT_SHARD_PRIMARY" - | "ETCD_TRANSACTION_FAILED" - | "SHARD_STOPPED"; + | 'OK' + | 'SHARD_NOT_FOUND' + | 'NO_SHARD_PRIMARY' + | 'NOT_SHARD_PRIMARY' + | 'ETCD_TRANSACTION_FAILED' + | 'SHARD_STOPPED'; export interface ConsumerUnassignResponse { - /** Status of the Unassign RPC. */ - status?: ConsumerStatus; - /** Shards which had assignments removed. */ - shards?: string[]; + /** Status of the Unassign RPC. */ + status?: ConsumerStatus; + /** Shards which had assignments removed. */ + shards?: string[]; } export interface ProtobufAny { - typeUrl?: string; - /** @format byte */ - value?: string; + typeUrl?: string; + /** @format byte */ + value?: string; } /** - * Header captures metadata such as the process responsible for processing - * an RPC, and its effective Etcd state. - */ +* Header captures metadata such as the process responsible for processing +an RPC, and its effective Etcd state. +*/ export interface ProtocolHeader { - /** - * ID of the process responsible for request processing. May be empty iff - * Header is being used within a proxied request, and that request may be - * dispatched to any member of the Route. - */ - processId?: ProcessSpecID; - /** - * Route of processes specifically responsible for this RPC, or an empty Route - * if any process is capable of serving the RPC. - */ - route?: ProtocolRoute; - /** - * Etcd represents the effective Etcd MVCC state under which a Gazette broker - * is operating in its processing of requests and responses. Its inclusion - * allows brokers to reason about relative "happened before" Revision ordering - * of apparent routing conflicts in proxied or replicated requests, as well - * as enabling sanity checks over equality of Etcd ClusterId (and precluding, - * for example, split-brain scenarios where different brokers are backed by - * different Etcd clusters). Etcd is kept in sync with - * etcdserverpb.ResponseHeader. - */ - etcd?: HeaderEtcd; + /** + * ID of the process responsible for request processing. May be empty iff + * Header is being used within a proxied request, and that request may be + * dispatched to any member of the Route. + */ + processId?: ProcessSpecID; + /** + * Route of processes specifically responsible for this RPC, or an empty Route + * if any process is capable of serving the RPC. + */ + route?: ProtocolRoute; + /** + * Etcd represents the effective Etcd MVCC state under which a Gazette broker + * is operating in its processing of requests and responses. Its inclusion + * allows brokers to reason about relative "happened before" Revision ordering + * of apparent routing conflicts in proxied or replicated requests, as well + * as enabling sanity checks over equality of Etcd ClusterId (and precluding, + * for example, split-brain scenarios where different brokers are backed by + * different Etcd clusters). Etcd is kept in sync with + * etcdserverpb.ResponseHeader. + */ + etcd?: HeaderEtcd; } /** - * Label defines a key & value pair which can be attached to entities like - * JournalSpecs and BrokerSpecs. Labels may be used to provide identifying - * attributes which do not directly imply semantics to the core system, but - * are meaningful to users or for higher-level Gazette tools. - */ +* Label defines a key & value pair which can be attached to entities like +JournalSpecs and BrokerSpecs. Labels may be used to provide identifying +attributes which do not directly imply semantics to the core system, but +are meaningful to users or for higher-level Gazette tools. +*/ export interface ProtocolLabel { - name?: string; - value?: string; + name?: string; + value?: string; } -/** LabelSelector defines a filter over LabelSets. */ +/** + * LabelSelector defines a filter over LabelSets. + */ export interface ProtocolLabelSelector { - /** - * Include is Labels which must be matched for a LabelSet to be selected. If - * empty, all Labels are included. An include Label with empty ("") value is - * matched by a Label of the same name having any value. - */ - include?: ProtocolLabelSet; - /** - * Exclude is Labels which cannot be matched for a LabelSet to be selected. If - * empty, no Labels are excluded. An exclude Label with empty ("") value - * excludes a Label of the same name having any value. - */ - exclude?: ProtocolLabelSet; + /** + * Include is Labels which must be matched for a LabelSet to be selected. If + * empty, all Labels are included. An include Label with empty ("") value is + * matched by a Label of the same name having any value. + */ + include?: ProtocolLabelSet; + /** + * Exclude is Labels which cannot be matched for a LabelSet to be selected. If + * empty, no Labels are excluded. An exclude Label with empty ("") value + * excludes a Label of the same name having any value. + */ + exclude?: ProtocolLabelSet; } -/** LabelSet is a collection of labels and their values. */ +/** + * LabelSet is a collection of labels and their values. + */ export interface ProtocolLabelSet { - /** Labels of the set. Instances must be unique and sorted over (Name, Value). */ - labels?: ProtocolLabel[]; + /** Labels of the set. Instances must be unique and sorted over (Name, Value). */ + labels?: ProtocolLabel[]; } -/** Route captures the current topology of an item and the processes serving it. */ +/** + * Route captures the current topology of an item and the processes serving it. + */ export interface ProtocolRoute { - /** Members of the Route, ordered on ascending ProcessSpec.ID (zone, suffix). */ - members?: ProcessSpecID[]; - /** - * Index of the ProcessSpec serving as primary within |members|, - * or -1 of no member is currently primary. - * @format int32 - */ - primary?: number; - /** - * Endpoints of each Route member. If not empty, |endpoints| has the same - * length and order as |members|, and captures the endpoint of each one. - */ - endpoints?: string[]; + /** Members of the Route, ordered on ascending ProcessSpec.ID (zone, suffix). */ + members?: ProcessSpecID[]; + /** + * Index of the ProcessSpec serving as primary within |members|, + * or -1 of no member is currently primary. + * @format int32 + */ + primary?: number; + /** + * Endpoints of each Route member. If not empty, |endpoints| has the same + * length and order as |members|, and captures the endpoint of each one. + */ + endpoints?: string[]; } /** - * FSMHints represents a manifest of Fnodes which were still live (eg, having - * remaining links) at the time the FSMHints were produced, as well as any - * Properties. It allows a Player of the log to identify minimal Segments which - * must be read to recover all Fnodes, and also contains sufficient metadata for - * a Player to resolve all possible conflicts it could encounter while reading - * the log, to arrive at a consistent view of file state which exactly matches - * that of the Recorder producing the FSMHints. - * Next tag: 4. - */ +* FSMHints represents a manifest of Fnodes which were still live (eg, having +remaining links) at the time the FSMHints were produced, as well as any +Properties. It allows a Player of the log to identify minimal Segments which +must be read to recover all Fnodes, and also contains sufficient metadata for +a Player to resolve all possible conflicts it could encounter while reading +the log, to arrive at a consistent view of file state which exactly matches +that of the Recorder producing the FSMHints. +Next tag: 4. +*/ export interface RecoverylogFSMHints { - /** - * Log is the implied recovery log of any contained |live_nodes| Segments - * which omit a |log| value. This implied behavior is both for backward- - * compatibility (Segments didn't always have a |log| field) and also for - * compacting the representation in the common case of Segments mostly or - * entirely addressing a single log. - */ - log?: string; - /** Live Fnodes and their Segments as-of the generation of these FSMHints. */ - liveNodes?: RecoverylogFnodeSegments[]; - /** Property files and contents as-of the generation of these FSMHints. */ - properties?: RecoverylogProperty[]; + /** + * Log is the implied recovery log of any contained |live_nodes| Segments + * which omit a |log| value. This implied behavior is both for backward- + * compatibility (Segments didn't always have a |log| field) and also for + * compacting the representation in the common case of Segments mostly or + * entirely addressing a single log. + */ + log?: string; + /** Live Fnodes and their Segments as-of the generation of these FSMHints. */ + liveNodes?: RecoverylogFnodeSegments[]; + /** Property files and contents as-of the generation of these FSMHints. */ + properties?: RecoverylogProperty[]; } -/** FnodeSegments captures log Segments containing all RecordedOps of the Fnode. */ +/** + * FnodeSegments captures log Segments containing all RecordedOps of the Fnode. + */ export interface RecoverylogFnodeSegments { - /** - * Fnode being hinted. - * @format int64 - */ - fnode?: string; - /** - * Segments of the Fnode in the log. Currently, FSM tracks only a single - * Segment per Fnode per Author & Log. A specific implication of this is that Fnodes - * modified over long periods of time will result in Segments spanning large - * chunks of the log. For best performance, Fnodes should be opened & written - * once, and then never be modified again (this is RocksDB's behavior). - * If supporting this case is desired, FSM will have to be a bit smarter about - * not extending Segments which gap over significant portions of the log - * (eg, there's a trade-off to make over size of the hinted manifest, vs - * savings incurred on playback by being able to skip portions of the log). - */ - segments?: RecoverylogSegment[]; + /** + * Fnode being hinted. + * @format int64 + */ + fnode?: string; + /** + * Segments of the Fnode in the log. Currently, FSM tracks only a single + * Segment per Fnode per Author & Log. A specific implication of this is that Fnodes + * modified over long periods of time will result in Segments spanning large + * chunks of the log. For best performance, Fnodes should be opened & written + * once, and then never be modified again (this is RocksDB's behavior). + * If supporting this case is desired, FSM will have to be a bit smarter about + * not extending Segments which gap over significant portions of the log + * (eg, there's a trade-off to make over size of the hinted manifest, vs + * savings incurred on playback by being able to skip portions of the log). + */ + segments?: RecoverylogSegment[]; } /** - * Property is a small file which rarely changes, and is thus managed - * outside of regular Fnode tracking. See FSM.Properties. - */ +* Property is a small file which rarely changes, and is thus managed +outside of regular Fnode tracking. See FSM.Properties. +*/ export interface RecoverylogProperty { - /** Filesystem path of this property, relative to the common base directory. */ - path?: string; - /** Complete file content of this property. */ - content?: string; + /** Filesystem path of this property, relative to the common base directory. */ + path?: string; + /** Complete file content of this property. */ + content?: string; } /** - * Segment is a contiguous chunk of recovery log written by a single Author. - * Recorders track Segments they have written, for use in providing hints to - * future readers of the log. A key point to understand is that Gazette append - * semantics mean that Recorders *cannot know* exactly what offsets their writes - * are applied to in the log, nor guarantee that their operations are not being - * interleaved with those of other writers. Log Players are aware of these - * limitations, and use Segments to resolve conflicts of possible interpretation - * of the log. Segments produced by a Player are exact, since Players observe all - * recorded operations at their exact offsets. - * Next tag: 8. - */ +* Segment is a contiguous chunk of recovery log written by a single Author. +Recorders track Segments they have written, for use in providing hints to +future readers of the log. A key point to understand is that Gazette append +semantics mean that Recorders *cannot know* exactly what offsets their writes +are applied to in the log, nor guarantee that their operations are not being +interleaved with those of other writers. Log Players are aware of these +limitations, and use Segments to resolve conflicts of possible interpretation +of the log. Segments produced by a Player are exact, since Players observe all +recorded operations at their exact offsets. +Next tag: 8. +*/ export interface RecoverylogSegment { - /** - * Author which wrote RecordedOps of this Segment. - * @format int64 - */ - author?: number; - /** - * First (lowest) sequence number of RecordedOps within this Segment. - * @format int64 - */ - firstSeqNo?: string; - /** - * First byte offset of the Segment, where |first_seq_no| is recorded. - * If this Segment was produced by a Recorder, this is guaranteed only to be a - * lower-bound (eg, a Player reading at this offset may encounter irrelevant - * operations prior to the RecordedOp indicated by the tuple - * (|author|, |first_seq_no|, |first_checksum|). If a Player produced the Segment, - * first_offset is exact. - * @format int64 - */ - firstOffset?: string; - /** - * Checksum of the RecordedOp having |first_seq_no|. - * @format int64 - */ - firstChecksum?: number; - /** - * Last (highest, inclusive) sequence number of RecordedOps within this Segment. - * @format int64 - */ - lastSeqNo?: string; - /** - * Last offset (exclusive) of the Segment. Zero means the offset is not known - * (eg, because the Segment was produced by a Recorder). - * @format int64 - */ - lastOffset?: string; - /** Log is the Journal holding this Segment's data, and to which offsets are relative. */ - log?: string; + /** + * Author which wrote RecordedOps of this Segment. + * @format int64 + */ + author?: number; + /** + * First (lowest) sequence number of RecordedOps within this Segment. + * @format int64 + */ + firstSeqNo?: string; + /** + * First byte offset of the Segment, where |first_seq_no| is recorded. + * If this Segment was produced by a Recorder, this is guaranteed only to be a + * lower-bound (eg, a Player reading at this offset may encounter irrelevant + * operations prior to the RecordedOp indicated by the tuple + * (|author|, |first_seq_no|, |first_checksum|). If a Player produced the Segment, + * first_offset is exact. + * @format int64 + */ + firstOffset?: string; + /** + * Checksum of the RecordedOp having |first_seq_no|. + * @format int64 + */ + firstChecksum?: number; + /** + * Last (highest, inclusive) sequence number of RecordedOps within this Segment. + * @format int64 + */ + lastSeqNo?: string; + /** + * Last offset (exclusive) of the Segment. Zero means the offset is not known + * (eg, because the Segment was produced by a Recorder). + * @format int64 + */ + lastOffset?: string; + /** Log is the Journal holding this Segment's data, and to which offsets are relative. */ + log?: string; } export interface RuntimeError { - error?: string; - /** @format int32 */ - code?: number; - message?: string; - details?: ProtobufAny[]; + error?: string; + /** @format int32 */ + code?: number; + message?: string; + details?: ProtobufAny[]; } diff --git a/client/dist/types/gen/broker/protocol/broker.d.ts b/client/dist/types/gen/broker/protocol/broker.d.ts index aefd4dc..66d4208 100644 --- a/client/dist/types/gen/broker/protocol/broker.d.ts +++ b/client/dist/types/gen/broker/protocol/broker.d.ts @@ -1,7 +1,7 @@ /** - * Change defines an insertion, update, or deletion to be applied to the set - * of JournalSpecs. Exactly one of |upsert| or |delete| must be set. - */ +* Change defines an insertion, update, or deletion to be applied to the set +of JournalSpecs. Exactly one of |upsert| or |delete| must be set. +*/ export interface ApplyRequestChange { /** * Expected ModRevision of the current JournalSpec. If the Journal is being @@ -17,7 +17,9 @@ export interface ApplyRequestChange { /** Journal to be deleted. expect_mod_revision must not be zero. */ delete?: string; } -/** Fragments of the Response. */ +/** + * Fragments of the Response. + */ export interface FragmentsResponseFragment { /** * Fragment is a content-addressed description of a contiguous Journal span, @@ -33,15 +35,15 @@ export interface FragmentsResponseFragment { signedUrl?: string; } /** - * Etcd represents the effective Etcd MVCC state under which a Gazette broker - * is operating in its processing of requests and responses. Its inclusion - * allows brokers to reason about relative "happened before" Revision ordering - * of apparent routing conflicts in proxied or replicated requests, as well - * as enabling sanity checks over equality of Etcd ClusterId (and precluding, - * for example, split-brain scenarios where different brokers are backed by - * different Etcd clusters). Etcd is kept in sync with - * etcdserverpb.ResponseHeader. - */ +* Etcd represents the effective Etcd MVCC state under which a Gazette broker +is operating in its processing of requests and responses. Its inclusion +allows brokers to reason about relative "happened before" Revision ordering +of apparent routing conflicts in proxied or replicated requests, as well +as enabling sanity checks over equality of Etcd ClusterId (and precluding, +for example, split-brain scenarios where different brokers are backed by +different Etcd clusters). Etcd is kept in sync with +etcdserverpb.ResponseHeader. +*/ export interface HeaderEtcd { /** * cluster_id is the ID of the cluster. @@ -65,7 +67,9 @@ export interface HeaderEtcd { */ raftTerm?: string; } -/** Journals of the response. */ +/** + * Journals of the response. + */ export interface ListResponseJournal { /** JournalSpec describes a Journal and its configuration. */ spec?: ProtocolJournalSpec; @@ -77,7 +81,9 @@ export interface ListResponseJournal { /** Route of the journal, including endpoints. */ route?: ProtocolRoute; } -/** ID composes a zone and a suffix to uniquely identify a ProcessSpec. */ +/** + * ID composes a zone and a suffix to uniquely identify a ProcessSpec. + */ export interface ProcessSpecID { /** * "Zone" in which the process is running. Zones may be AWS, Azure, or @@ -100,7 +106,9 @@ export interface ProtobufAny { /** @format byte */ value?: string; } -/** AppendResponse is the unary response message of the broker Append RPC. */ +/** + * AppendResponse is the unary response message of the broker Append RPC. + */ export interface ProtocolAppendResponse { /** Status of the Append RPC. */ status?: ProtocolStatus; @@ -124,7 +132,9 @@ export interface ProtocolAppendResponse { */ delayedChunks?: string; } -/** ApplyResponse is the unary response message of the broker Apply RPC. */ +/** + * ApplyResponse is the unary response message of the broker Apply RPC. + */ export interface ProtocolApplyResponse { /** Status of the Apply RPC. */ status?: ProtocolStatus; @@ -132,36 +142,35 @@ export interface ProtocolApplyResponse { header?: ProtocolHeader; } /** - * CompressionCode defines codecs known to Gazette. - * - * - INVALID: INVALID is the zero-valued CompressionCodec, and is not a valid codec. - * - NONE: NONE encodes Fragments without any applied compression, with default suffix - * ".raw". - * - GZIP: GZIP encodes Fragments using the Gzip library, with default suffix ".gz". - * - ZSTANDARD: ZSTANDARD encodes Fragments using the ZStandard library, with default - * suffix ".zst". - * - SNAPPY: SNAPPY encodes Fragments using the Snappy library, with default suffix - * ".sz". - * - GZIP_OFFLOAD_DECOMPRESSION: GZIP_OFFLOAD_DECOMPRESSION is the GZIP codec with additional behavior - * around reads and writes to remote Fragment stores, designed to offload - * the work of decompression onto compatible stores. Specifically: - * * Fragments are written with a "Content-Encoding: gzip" header. - * * Client read requests are made with "Accept-Encoding: identity". - * This can be helpful in contexts where reader IO bandwidth to the storage - * API is unconstrained, as the cost of decompression is offloaded to the - * store and CPU-intensive batch readers may receive a parallelism benefit. - * While this codec may provide substantial read-time performance - * improvements, it is an advanced configuration and the "Content-Encoding" - * header handling can be subtle and sometimes confusing. It uses the default - * suffix ".gzod". - * @default "INVALID" - */ -export declare type ProtocolCompressionCodec = "INVALID" | "NONE" | "GZIP" | "ZSTANDARD" | "SNAPPY" | "GZIP_OFFLOAD_DECOMPRESSION"; +* CompressionCode defines codecs known to Gazette. + + - INVALID: INVALID is the zero-valued CompressionCodec, and is not a valid codec. + - NONE: NONE encodes Fragments without any applied compression, with default suffix +".raw". + - GZIP: GZIP encodes Fragments using the Gzip library, with default suffix ".gz". + - ZSTANDARD: ZSTANDARD encodes Fragments using the ZStandard library, with default +suffix ".zst". + - SNAPPY: SNAPPY encodes Fragments using the Snappy library, with default suffix +".sz". + - GZIP_OFFLOAD_DECOMPRESSION: GZIP_OFFLOAD_DECOMPRESSION is the GZIP codec with additional behavior +around reads and writes to remote Fragment stores, designed to offload +the work of decompression onto compatible stores. Specifically: + * Fragments are written with a "Content-Encoding: gzip" header. + * Client read requests are made with "Accept-Encoding: identity". +This can be helpful in contexts where reader IO bandwidth to the storage +API is unconstrained, as the cost of decompression is offloaded to the +store and CPU-intensive batch readers may receive a parallelism benefit. +While this codec may provide substantial read-time performance +improvements, it is an advanced configuration and the "Content-Encoding" +header handling can be subtle and sometimes confusing. It uses the default +suffix ".gzod". +*/ +export declare type ProtocolCompressionCodec = 'INVALID' | 'NONE' | 'GZIP' | 'ZSTANDARD' | 'SNAPPY' | 'GZIP_OFFLOAD_DECOMPRESSION'; /** - * Fragment is a content-addressed description of a contiguous Journal span, - * defined by the [begin, end) offset range covered by the Fragment and the - * SHA1 sum of the corresponding Journal content. - */ +* Fragment is a content-addressed description of a contiguous Journal span, +defined by the [begin, end) offset range covered by the Fragment and the +SHA1 sum of the corresponding Journal content. +*/ export interface ProtocolFragment { /** Journal of the Fragment. */ journal?: string; @@ -196,9 +205,9 @@ export interface ProtocolFragment { pathPostfix?: string; } /** - * FragmentsResponse is the unary response message of the broker ListFragments - * RPC. - */ +* FragmentsResponse is the unary response message of the broker ListFragments +RPC. +*/ export interface ProtocolFragmentsResponse { /** Status of the Apply RPC. */ status?: ProtocolStatus; @@ -214,9 +223,9 @@ export interface ProtocolFragmentsResponse { nextPageToken?: string; } /** - * Header captures metadata such as the process responsible for processing - * an RPC, and its effective Etcd state. - */ +* Header captures metadata such as the process responsible for processing +an RPC, and its effective Etcd state. +*/ export interface ProtocolHeader { /** * ID of the process responsible for request processing. May be empty iff @@ -241,7 +250,9 @@ export interface ProtocolHeader { */ etcd?: HeaderEtcd; } -/** JournalSpec describes a Journal and its configuration. */ +/** + * JournalSpec describes a Journal and its configuration. + */ export interface ProtocolJournalSpec { /** Name of the Journal. */ name?: string; @@ -278,9 +289,9 @@ export interface ProtocolJournalSpec { maxAppendRate?: string; } /** - * Fragment is JournalSpec configuration which pertains to the creation, - * persistence, and indexing of the Journal's Fragments. - */ +* Fragment is JournalSpec configuration which pertains to the creation, +persistence, and indexing of the Journal's Fragments. +*/ export interface ProtocolJournalSpecFragment { /** * Target content length of each Fragment. In normal operation after @@ -338,7 +349,6 @@ export interface ProtocolJournalSpecFragment { * fragment on the hour, every hour, even if it has not yet reached its * target length. A batch work-flow running at 5 minutes past the hour is * then reasonably assured of seeing all events from the past hour. - * * See also "gazctl journals fragments --help" for more discussion. */ flushInterval?: string; @@ -359,16 +369,18 @@ export interface ProtocolJournalSpecFragment { pathPostfixTemplate?: string; } /** - * Label defines a key & value pair which can be attached to entities like - * JournalSpecs and BrokerSpecs. Labels may be used to provide identifying - * attributes which do not directly imply semantics to the core system, but - * are meaningful to users or for higher-level Gazette tools. - */ +* Label defines a key & value pair which can be attached to entities like +JournalSpecs and BrokerSpecs. Labels may be used to provide identifying +attributes which do not directly imply semantics to the core system, but +are meaningful to users or for higher-level Gazette tools. +*/ export interface ProtocolLabel { name?: string; value?: string; } -/** LabelSelector defines a filter over LabelSets. */ +/** + * LabelSelector defines a filter over LabelSets. + */ export interface ProtocolLabelSelector { /** * Include is Labels which must be matched for a LabelSet to be selected. If @@ -383,12 +395,16 @@ export interface ProtocolLabelSelector { */ exclude?: ProtocolLabelSet; } -/** LabelSet is a collection of labels and their values. */ +/** + * LabelSet is a collection of labels and their values. + */ export interface ProtocolLabelSet { /** Labels of the set. Instances must be unique and sorted over (Name, Value). */ labels?: ProtocolLabel[]; } -/** ListRequest is the unary request message of the broker List RPC. */ +/** + * ListRequest is the unary request message of the broker List RPC. + */ export interface ProtocolListRequest { /** * Selector optionally refines the set of journals which will be enumerated. @@ -402,7 +418,9 @@ export interface ProtocolListRequest { */ selector?: ProtocolLabelSelector; } -/** ListResponse is the unary response message of the broker List RPC. */ +/** + * ListResponse is the unary response message of the broker List RPC. + */ export interface ProtocolListResponse { /** Status of the List RPC. */ status?: ProtocolStatus; @@ -410,7 +428,9 @@ export interface ProtocolListResponse { header?: ProtocolHeader; journals?: ListResponseJournal[]; } -/** ReadRequest is the unary request message of the broker Read RPC. */ +/** + * ReadRequest is the unary request message of the broker Read RPC. + */ export interface ProtocolReadRequest { /** Header is attached by a proxying broker peer. */ header?: ProtocolHeader; @@ -447,21 +467,29 @@ export interface ProtocolReadRequest { * @format int64 */ endOffset?: string; + /** + * BeginModTime is an optional inclusive lower bound on the modification + * timestamps of fragments read from the backing store, represented as + * seconds since the epoch. The request Offset will be advanced as-needed + * to skip persisted Fragments having a modication time before the bound. + * @format int64 + */ + beginModTime?: string; } /** - * ReadResponse is the streamed response message of the broker Read RPC. - * Responses messages are of two types: - * * "Metadata" messages, which conveys the journal Fragment addressed by the - * request which is ready to be read. - * * "Chunk" messages, which carry associated journal Fragment content bytes. - * - * A metadata message specifying a Fragment always precedes all "chunks" of the - * Fragment's content. Response streams may be very long lived, having many - * metadata and accompanying chunk messages. The reader may also block for long - * periods of time awaiting the next metadata message (eg, if the next offset - * hasn't yet committed). However once a metadata message is read, the reader - * is assured that its associated chunk messages are immediately forthcoming. - */ +* ReadResponse is the streamed response message of the broker Read RPC. +Responses messages are of two types: +* * "Metadata" messages, which conveys the journal Fragment addressed by the + request which is ready to be read. +* "Chunk" messages, which carry associated journal Fragment content bytes. + +A metadata message specifying a Fragment always precedes all "chunks" of the +Fragment's content. Response streams may be very long lived, having many +metadata and accompanying chunk messages. The reader may also block for long +periods of time awaiting the next metadata message (eg, if the next offset +hasn't yet committed). However once a metadata message is read, the reader +is assured that its associated chunk messages are immediately forthcoming. +*/ export interface ProtocolReadResponse { /** Status of the Read RPC. */ status?: ProtocolStatus; @@ -497,10 +525,10 @@ export interface ProtocolReadResponse { content?: string; } /** - * ReplicateResponse is the streamed response message of the broker's internal - * Replicate RPC. Each message is a 1:1 response to a previously read "proposal" - * ReplicateRequest with |acknowledge| set. - */ +* ReplicateResponse is the streamed response message of the broker's internal +Replicate RPC. Each message is a 1:1 response to a previously read "proposal" +ReplicateRequest with |acknowledge| set. +*/ export interface ProtocolReplicateResponse { /** Status of the Replicate RPC. */ status?: ProtocolStatus; @@ -521,7 +549,9 @@ export interface ProtocolReplicateResponse { */ registers?: ProtocolLabelSet; } -/** Route captures the current topology of an item and the processes serving it. */ +/** + * Route captures the current topology of an item and the processes serving it. + */ export interface ProtocolRoute { /** Members of the Route, ordered on ascending ProcessSpec.ID (zone, suffix). */ members?: ProcessSpecID[]; @@ -537,7 +567,9 @@ export interface ProtocolRoute { */ endpoints?: string[]; } -/** SHA1Sum is a 160-bit SHA1 digest. */ +/** + * SHA1Sum is a 160-bit SHA1 digest. + */ export interface ProtocolSHA1Sum { /** @format uint64 */ part1?: string; @@ -547,37 +579,36 @@ export interface ProtocolSHA1Sum { part3?: number; } /** - * Status is a response status code, used universally across Gazette RPC APIs. - * - * - JOURNAL_NOT_FOUND: The named journal does not exist. - * - NO_JOURNAL_PRIMARY_BROKER: There is no current primary broker for the journal. This is a temporary - * condition which should quickly resolve, assuming sufficient broker - * capacity. - * - NOT_JOURNAL_PRIMARY_BROKER: The present broker is not the assigned primary broker for the journal. - * - NOT_JOURNAL_BROKER: The present broker is not an assigned broker for the journal. - * - INSUFFICIENT_JOURNAL_BROKERS: There are an insufficient number of assigned brokers for the journal - * to meet its required replication. - * - OFFSET_NOT_YET_AVAILABLE: The requested offset is not yet available. This indicates either that the - * offset has not yet been written, or that the broker is not yet aware of a - * written fragment covering the offset. Returned only by non-blocking reads. - * - WRONG_ROUTE: The peer disagrees with the Route accompanying a ReplicateRequest. - * - PROPOSAL_MISMATCH: The peer disagrees with the proposal accompanying a ReplicateRequest. - * - ETCD_TRANSACTION_FAILED: The Etcd transaction failed. Returned by Update RPC when an - * expect_mod_revision of the UpdateRequest differs from the current - * ModRevision of the JournalSpec within the store. - * - NOT_ALLOWED: A disallowed journal access was attempted (eg, a write where the - * journal disables writes, or read where journals disable reads). - * - WRONG_APPEND_OFFSET: The Append is refused because its requested offset is not equal - * to the furthest written offset of the journal. - * - INDEX_HAS_GREATER_OFFSET: The Append is refused because the replication pipeline tracks a smaller - * journal offset than that of the remote fragment index. This indicates - * that journal replication consistency has been lost in the past, due to - * too many broker or Etcd failures. - * - REGISTER_MISMATCH: The Append is refused because a registers selector was provided with the - * request, but it was not matched by current register values of the journal. - * @default "OK" - */ -export declare type ProtocolStatus = "OK" | "JOURNAL_NOT_FOUND" | "NO_JOURNAL_PRIMARY_BROKER" | "NOT_JOURNAL_PRIMARY_BROKER" | "NOT_JOURNAL_BROKER" | "INSUFFICIENT_JOURNAL_BROKERS" | "OFFSET_NOT_YET_AVAILABLE" | "WRONG_ROUTE" | "PROPOSAL_MISMATCH" | "ETCD_TRANSACTION_FAILED" | "NOT_ALLOWED" | "WRONG_APPEND_OFFSET" | "INDEX_HAS_GREATER_OFFSET" | "REGISTER_MISMATCH"; +* Status is a response status code, used universally across Gazette RPC APIs. + + - JOURNAL_NOT_FOUND: The named journal does not exist. + - NO_JOURNAL_PRIMARY_BROKER: There is no current primary broker for the journal. This is a temporary +condition which should quickly resolve, assuming sufficient broker +capacity. + - NOT_JOURNAL_PRIMARY_BROKER: The present broker is not the assigned primary broker for the journal. + - NOT_JOURNAL_BROKER: The present broker is not an assigned broker for the journal. + - INSUFFICIENT_JOURNAL_BROKERS: There are an insufficient number of assigned brokers for the journal +to meet its required replication. + - OFFSET_NOT_YET_AVAILABLE: The requested offset is not yet available. This indicates either that the +offset has not yet been written, or that the broker is not yet aware of a +written fragment covering the offset. Returned only by non-blocking reads. + - WRONG_ROUTE: The peer disagrees with the Route accompanying a ReplicateRequest. + - PROPOSAL_MISMATCH: The peer disagrees with the proposal accompanying a ReplicateRequest. + - ETCD_TRANSACTION_FAILED: The Etcd transaction failed. Returned by Update RPC when an +expect_mod_revision of the UpdateRequest differs from the current +ModRevision of the JournalSpec within the store. + - NOT_ALLOWED: A disallowed journal access was attempted (eg, a write where the +journal disables writes, or read where journals disable reads). + - WRONG_APPEND_OFFSET: The Append is refused because its requested offset is not equal +to the furthest written offset of the journal. + - INDEX_HAS_GREATER_OFFSET: The Append is refused because the replication pipeline tracks a smaller +journal offset than that of the remote fragment index. This indicates +that journal replication consistency has been lost in the past, due to +too many broker or Etcd failures. + - REGISTER_MISMATCH: The Append is refused because a registers selector was provided with the +request, but it was not matched by current register values of the journal. +*/ +export declare type ProtocolStatus = 'OK' | 'JOURNAL_NOT_FOUND' | 'NO_JOURNAL_PRIMARY_BROKER' | 'NOT_JOURNAL_PRIMARY_BROKER' | 'NOT_JOURNAL_BROKER' | 'INSUFFICIENT_JOURNAL_BROKERS' | 'OFFSET_NOT_YET_AVAILABLE' | 'WRONG_ROUTE' | 'PROPOSAL_MISMATCH' | 'ETCD_TRANSACTION_FAILED' | 'NOT_ALLOWED' | 'WRONG_APPEND_OFFSET' | 'INDEX_HAS_GREATER_OFFSET' | 'REGISTER_MISMATCH'; export interface RuntimeError { error?: string; /** @format int32 */ diff --git a/client/dist/types/gen/consumer/protocol/consumer.d.ts b/client/dist/types/gen/consumer/protocol/consumer.d.ts index 03dc0e8..59f5665 100644 --- a/client/dist/types/gen/consumer/protocol/consumer.d.ts +++ b/client/dist/types/gen/consumer/protocol/consumer.d.ts @@ -3,15 +3,15 @@ export interface GetHintsResponseResponseHints { hints?: RecoverylogFSMHints; } /** - * Etcd represents the effective Etcd MVCC state under which a Gazette broker - * is operating in its processing of requests and responses. Its inclusion - * allows brokers to reason about relative "happened before" Revision ordering - * of apparent routing conflicts in proxied or replicated requests, as well - * as enabling sanity checks over equality of Etcd ClusterId (and precluding, - * for example, split-brain scenarios where different brokers are backed by - * different Etcd clusters). Etcd is kept in sync with - * etcdserverpb.ResponseHeader. - */ +* Etcd represents the effective Etcd MVCC state under which a Gazette broker +is operating in its processing of requests and responses. Its inclusion +allows brokers to reason about relative "happened before" Revision ordering +of apparent routing conflicts in proxied or replicated requests, as well +as enabling sanity checks over equality of Etcd ClusterId (and precluding, +for example, split-brain scenarios where different brokers are backed by +different Etcd clusters). Etcd is kept in sync with +etcdserverpb.ResponseHeader. +*/ export interface HeaderEtcd { /** * cluster_id is the ID of the cluster. @@ -35,7 +35,9 @@ export interface HeaderEtcd { */ raftTerm?: string; } -/** Shards of the response. */ +/** + * Shards of the response. + */ export interface ListResponseShard { /** * ShardSpec describes a shard and its configuration, and is the long-lived unit @@ -60,7 +62,9 @@ export interface ListResponseShard { /** Status of each replica. Cardinality and ordering matches |route|. */ status?: ConsumerReplicaStatus[]; } -/** ID composes a zone and a suffix to uniquely identify a ProcessSpec. */ +/** + * ID composes a zone and a suffix to uniquely identify a ProcessSpec. + */ export interface ProcessSpecID { /** * "Zone" in which the process is running. Zones may be AWS, Azure, or @@ -79,21 +83,20 @@ export interface ProcessSpecID { suffix?: string; } /** - * - BACKFILL: The replica is actively playing the historical recovery log. - * - STANDBY: The replica has finished playing the historical recovery log and is - * live-tailing it to locally mirror recorded operations as they are - * produced. It can take over as primary at any time. - * - * Shards not having recovery logs immediately transition to STANDBY. - * - PRIMARY: The replica is actively serving as primary. - * - FAILED: The replica has encountered an unrecoverable error. - * @default "IDLE" - */ -export declare type ReplicaStatusCode = "IDLE" | "BACKFILL" | "STANDBY" | "PRIMARY" | "FAILED"; +* - BACKFILL: The replica is actively playing the historical recovery log. + - STANDBY: The replica has finished playing the historical recovery log and is +live-tailing it to locally mirror recorded operations as they are +produced. It can take over as primary at any time. + +Shards not having recovery logs immediately transition to STANDBY. + - PRIMARY: The replica is actively serving as primary. + - FAILED: The replica has encountered an unrecoverable error. +*/ +export declare type ReplicaStatusCode = 'IDLE' | 'BACKFILL' | 'STANDBY' | 'PRIMARY' | 'FAILED'; /** - * Change defines an insertion, update, or deletion to be applied to the set - * of ShardSpecs. Exactly one of |upsert| or |delete| must be set. - */ +* Change defines an insertion, update, or deletion to be applied to the set +of ShardSpecs. Exactly one of |upsert| or |delete| must be set. +*/ export interface ConsumerApplyRequestChange { /** * Expected ModRevision of the current ShardSpec. If the shard is being @@ -167,13 +170,13 @@ export interface ConsumerListResponse { extension?: string; } /** - * ReplicaStatus is the status of a ShardSpec assigned to a ConsumerSpec. - * It serves as an allocator AssignmentValue. ReplicaStatus is reduced by taking - * the maximum enum value among statuses. Eg, if a primary is PRIMARY, one - * replica is BACKFILL and the other STANDBY, then the status is PRIMARY. If one - * of the replicas transitioned to FAILED, than the status is FAILED. This - * reduction behavior is used to summarize status across all replicas. - */ +* ReplicaStatus is the status of a ShardSpec assigned to a ConsumerSpec. +It serves as an allocator AssignmentValue. ReplicaStatus is reduced by taking +the maximum enum value among statuses. Eg, if a primary is PRIMARY, one +replica is BACKFILL and the other STANDBY, then the status is PRIMARY. If one +of the replicas transitioned to FAILED, than the status is FAILED. This +reduction behavior is used to summarize status across all replicas. +*/ export interface ConsumerReplicaStatus { /** * - BACKFILL: The replica is actively playing the historical recovery log. @@ -190,17 +193,17 @@ export interface ConsumerReplicaStatus { errors?: string[]; } /** - * ShardSpec describes a shard and its configuration, and is the long-lived unit - * of work and scaling for a consumer application. Each shard is allocated to a - * one "primary" at-a-time selected from the current processes of a consumer - * application, and is re-assigned on process fault or exit. - * - * ShardSpecs describe all configuration of the shard and its processing, - * including journals to consume, configuration for processing transactions, its - * recovery log, hot standbys, etc. ShardSpecs may be further extended with - * domain-specific labels & values to further define application behavior. - * ShardSpec is-a allocator.ItemValue. - */ +* ShardSpec describes a shard and its configuration, and is the long-lived unit +of work and scaling for a consumer application. Each shard is allocated to a +one "primary" at-a-time selected from the current processes of a consumer +application, and is re-assigned on process fault or exit. + +ShardSpecs describe all configuration of the shard and its processing, +including journals to consume, configuration for processing transactions, its +recovery log, hot standbys, etc. ShardSpecs may be further extended with +domain-specific labels & values to further define application behavior. +ShardSpec is-a allocator.ItemValue. +*/ export interface ConsumerShardSpec { /** ID of the shard. */ id?: string; @@ -233,12 +236,10 @@ export interface ConsumerShardSpec { * "{hints_prefix/{shard_id}.backup.0" to * "{hints_prefix/{shard_id}.backup.1", and so on, keeping at most * |hint_backups| distinct sets of FSMHints. - * * In the case of disaster or data-loss, these copied hints can be an * important fallback for recovering a consistent albeit older version of the * shard's store, with each relying on only progressively older portions of * the recovery log. - * * When pruning the recovery log, log fragments which are older than (and no * longer required by) the *oldest* backup are discarded, ensuring that * all hints remain valid for playback. @@ -302,7 +303,6 @@ export interface ConsumerShardSpec { * forthcoming and, by remaining open, we can process all messages in this * transaction. Effectively we're trading a small amount of increased local * latency for a global reduction in end-to-end latency. - * * This works well for acyclic message flows, but can introduce unnecessary * stalls if there are message cycles between shards. In the simplest case, * a transaction could block awaiting an ACK of a message that it itself @@ -332,22 +332,22 @@ export interface ConsumerShardSpec { readChannelSize?: number; } /** - * Sources define the set of journals which this shard consumes. At least one - * Source must be specified, and in many use cases only one will be needed. - * For use cases which can benefit, multiple sources may be specified to - * represent a "join" over messages of distinct journals. - * - * Note the effective mapping of messages to each of the joined journals - * should align (eg, joining a journal of customer updates with one of orders, - * where both are mapped on customer ID). This typically means the - * partitioning of the two event "topics" must be the same. - * - * Another powerful pattern is to shard on partitions of a high-volume event - * stream, and also have each shard join against all events of a low-volume - * stream. For example, a shard might ingest and index "viewed product" - * events, read a comparably low-volume "purchase" event stream, and on each - * purchase publish the bundle of its corresponding prior product views. - */ +* Sources define the set of journals which this shard consumes. At least one +Source must be specified, and in many use cases only one will be needed. +For use cases which can benefit, multiple sources may be specified to +represent a "join" over messages of distinct journals. + +Note the effective mapping of messages to each of the joined journals +should align (eg, joining a journal of customer updates with one of orders, +where both are mapped on customer ID). This typically means the +partitioning of the two event "topics" must be the same. + +Another powerful pattern is to shard on partitions of a high-volume event +stream, and also have each shard join against all events of a low-volume +stream. For example, a shard might ingest and index "viewed product" +events, read a comparably low-volume "purchase" event stream, and on each +purchase publish the bundle of its corresponding prior product views. +*/ export interface ConsumerShardSpecSource { /** Journal which this shard is consuming. */ journal?: string; @@ -395,7 +395,6 @@ export interface ConsumerStatResponse { * Formally, if an acknowledged message A results in this shard publishing * messages B, and A falls within |read_through|, then all messages B & their * acknowledgements fall within |publish_at|. - * * The composition of |read_through| and |publish_at| allow CQRS applications * to provide read-your-writes consistency, even if written events pass * through multiple intermediate consumers and arbitrary transformations @@ -409,25 +408,24 @@ export interface ConsumerStatResponse { extension?: string; } /** - * Status is a response status code, used across Gazette Consumer RPC APIs. - * - * - SHARD_NOT_FOUND: The named shard does not exist. - * - NO_SHARD_PRIMARY: There is no current primary consumer process for the shard. This is a - * temporary condition which should quickly resolve, assuming sufficient - * consumer capacity. - * - NOT_SHARD_PRIMARY: The present consumer process is not the assigned primary for the shard, - * and was not instructed to proxy the request. - * - ETCD_TRANSACTION_FAILED: The Etcd transaction failed. Returned by Update RPC when an - * expect_mod_revision of the UpdateRequest differs from the current - * ModRevision of the ShardSpec within the store. - * - SHARD_STOPPED: The current primary shard has stopped, either due to reassignment or - * processing failure, and will not make further progress toward the - * requested operation. - * For example, a Stat RPC will return SHARD_STOPPED if the StatRequest - * cannot be satisfied. - * @default "OK" - */ -export declare type ConsumerStatus = "OK" | "SHARD_NOT_FOUND" | "NO_SHARD_PRIMARY" | "NOT_SHARD_PRIMARY" | "ETCD_TRANSACTION_FAILED" | "SHARD_STOPPED"; +* Status is a response status code, used across Gazette Consumer RPC APIs. + + - SHARD_NOT_FOUND: The named shard does not exist. + - NO_SHARD_PRIMARY: There is no current primary consumer process for the shard. This is a +temporary condition which should quickly resolve, assuming sufficient +consumer capacity. + - NOT_SHARD_PRIMARY: The present consumer process is not the assigned primary for the shard, +and was not instructed to proxy the request. + - ETCD_TRANSACTION_FAILED: The Etcd transaction failed. Returned by Update RPC when an +expect_mod_revision of the UpdateRequest differs from the current +ModRevision of the ShardSpec within the store. + - SHARD_STOPPED: The current primary shard has stopped, either due to reassignment or +processing failure, and will not make further progress toward the +requested operation. +For example, a Stat RPC will return SHARD_STOPPED if the StatRequest +cannot be satisfied. +*/ +export declare type ConsumerStatus = 'OK' | 'SHARD_NOT_FOUND' | 'NO_SHARD_PRIMARY' | 'NOT_SHARD_PRIMARY' | 'ETCD_TRANSACTION_FAILED' | 'SHARD_STOPPED'; export interface ConsumerUnassignResponse { /** Status of the Unassign RPC. */ status?: ConsumerStatus; @@ -440,9 +438,9 @@ export interface ProtobufAny { value?: string; } /** - * Header captures metadata such as the process responsible for processing - * an RPC, and its effective Etcd state. - */ +* Header captures metadata such as the process responsible for processing +an RPC, and its effective Etcd state. +*/ export interface ProtocolHeader { /** * ID of the process responsible for request processing. May be empty iff @@ -468,16 +466,18 @@ export interface ProtocolHeader { etcd?: HeaderEtcd; } /** - * Label defines a key & value pair which can be attached to entities like - * JournalSpecs and BrokerSpecs. Labels may be used to provide identifying - * attributes which do not directly imply semantics to the core system, but - * are meaningful to users or for higher-level Gazette tools. - */ +* Label defines a key & value pair which can be attached to entities like +JournalSpecs and BrokerSpecs. Labels may be used to provide identifying +attributes which do not directly imply semantics to the core system, but +are meaningful to users or for higher-level Gazette tools. +*/ export interface ProtocolLabel { name?: string; value?: string; } -/** LabelSelector defines a filter over LabelSets. */ +/** + * LabelSelector defines a filter over LabelSets. + */ export interface ProtocolLabelSelector { /** * Include is Labels which must be matched for a LabelSet to be selected. If @@ -492,12 +492,16 @@ export interface ProtocolLabelSelector { */ exclude?: ProtocolLabelSet; } -/** LabelSet is a collection of labels and their values. */ +/** + * LabelSet is a collection of labels and their values. + */ export interface ProtocolLabelSet { /** Labels of the set. Instances must be unique and sorted over (Name, Value). */ labels?: ProtocolLabel[]; } -/** Route captures the current topology of an item and the processes serving it. */ +/** + * Route captures the current topology of an item and the processes serving it. + */ export interface ProtocolRoute { /** Members of the Route, ordered on ascending ProcessSpec.ID (zone, suffix). */ members?: ProcessSpecID[]; @@ -514,15 +518,15 @@ export interface ProtocolRoute { endpoints?: string[]; } /** - * FSMHints represents a manifest of Fnodes which were still live (eg, having - * remaining links) at the time the FSMHints were produced, as well as any - * Properties. It allows a Player of the log to identify minimal Segments which - * must be read to recover all Fnodes, and also contains sufficient metadata for - * a Player to resolve all possible conflicts it could encounter while reading - * the log, to arrive at a consistent view of file state which exactly matches - * that of the Recorder producing the FSMHints. - * Next tag: 4. - */ +* FSMHints represents a manifest of Fnodes which were still live (eg, having +remaining links) at the time the FSMHints were produced, as well as any +Properties. It allows a Player of the log to identify minimal Segments which +must be read to recover all Fnodes, and also contains sufficient metadata for +a Player to resolve all possible conflicts it could encounter while reading +the log, to arrive at a consistent view of file state which exactly matches +that of the Recorder producing the FSMHints. +Next tag: 4. +*/ export interface RecoverylogFSMHints { /** * Log is the implied recovery log of any contained |live_nodes| Segments @@ -537,7 +541,9 @@ export interface RecoverylogFSMHints { /** Property files and contents as-of the generation of these FSMHints. */ properties?: RecoverylogProperty[]; } -/** FnodeSegments captures log Segments containing all RecordedOps of the Fnode. */ +/** + * FnodeSegments captures log Segments containing all RecordedOps of the Fnode. + */ export interface RecoverylogFnodeSegments { /** * Fnode being hinted. @@ -558,9 +564,9 @@ export interface RecoverylogFnodeSegments { segments?: RecoverylogSegment[]; } /** - * Property is a small file which rarely changes, and is thus managed - * outside of regular Fnode tracking. See FSM.Properties. - */ +* Property is a small file which rarely changes, and is thus managed +outside of regular Fnode tracking. See FSM.Properties. +*/ export interface RecoverylogProperty { /** Filesystem path of this property, relative to the common base directory. */ path?: string; @@ -568,17 +574,17 @@ export interface RecoverylogProperty { content?: string; } /** - * Segment is a contiguous chunk of recovery log written by a single Author. - * Recorders track Segments they have written, for use in providing hints to - * future readers of the log. A key point to understand is that Gazette append - * semantics mean that Recorders *cannot know* exactly what offsets their writes - * are applied to in the log, nor guarantee that their operations are not being - * interleaved with those of other writers. Log Players are aware of these - * limitations, and use Segments to resolve conflicts of possible interpretation - * of the log. Segments produced by a Player are exact, since Players observe all - * recorded operations at their exact offsets. - * Next tag: 8. - */ +* Segment is a contiguous chunk of recovery log written by a single Author. +Recorders track Segments they have written, for use in providing hints to +future readers of the log. A key point to understand is that Gazette append +semantics mean that Recorders *cannot know* exactly what offsets their writes +are applied to in the log, nor guarantee that their operations are not being +interleaved with those of other writers. Log Players are aware of these +limitations, and use Segments to resolve conflicts of possible interpretation +of the log. Segments produced by a Player are exact, since Players observe all +recorded operations at their exact offsets. +Next tag: 8. +*/ export interface RecoverylogSegment { /** * Author which wrote RecordedOps of this Segment. diff --git a/client/src/gen/broker/protocol/broker.ts b/client/src/gen/broker/protocol/broker.ts index b2366bc..964b6ea 100644 --- a/client/src/gen/broker/protocol/broker.ts +++ b/client/src/gen/broker/protocol/broker.ts @@ -10,644 +10,675 @@ */ /** - * Change defines an insertion, update, or deletion to be applied to the set - * of JournalSpecs. Exactly one of |upsert| or |delete| must be set. - */ +* Change defines an insertion, update, or deletion to be applied to the set +of JournalSpecs. Exactly one of |upsert| or |delete| must be set. +*/ export interface ApplyRequestChange { - /** - * Expected ModRevision of the current JournalSpec. If the Journal is being - * created, expect_mod_revision is zero. - * @format int64 - */ - expectModRevision?: string; - /** - * JournalSpec to be updated (if expect_mod_revision > 0) or created - * (if expect_mod_revision == 0). - */ - upsert?: ProtocolJournalSpec; - /** Journal to be deleted. expect_mod_revision must not be zero. */ - delete?: string; + /** + * Expected ModRevision of the current JournalSpec. If the Journal is being + * created, expect_mod_revision is zero. + * @format int64 + */ + expectModRevision?: string; + /** + * JournalSpec to be updated (if expect_mod_revision > 0) or created + * (if expect_mod_revision == 0). + */ + upsert?: ProtocolJournalSpec; + /** Journal to be deleted. expect_mod_revision must not be zero. */ + delete?: string; } -/** Fragments of the Response. */ +/** + * Fragments of the Response. + */ export interface FragmentsResponseFragment { - /** - * Fragment is a content-addressed description of a contiguous Journal span, - * defined by the [begin, end) offset range covered by the Fragment and the - * SHA1 sum of the corresponding Journal content. - */ - spec?: ProtocolFragment; - /** - * SignedURL is a temporary URL at which a direct GET of the Fragment may - * be issued, signed by the broker's credentials. Set only if the request - * specified a SignatureTTL. - */ - signedUrl?: string; + /** + * Fragment is a content-addressed description of a contiguous Journal span, + * defined by the [begin, end) offset range covered by the Fragment and the + * SHA1 sum of the corresponding Journal content. + */ + spec?: ProtocolFragment; + /** + * SignedURL is a temporary URL at which a direct GET of the Fragment may + * be issued, signed by the broker's credentials. Set only if the request + * specified a SignatureTTL. + */ + signedUrl?: string; } /** - * Etcd represents the effective Etcd MVCC state under which a Gazette broker - * is operating in its processing of requests and responses. Its inclusion - * allows brokers to reason about relative "happened before" Revision ordering - * of apparent routing conflicts in proxied or replicated requests, as well - * as enabling sanity checks over equality of Etcd ClusterId (and precluding, - * for example, split-brain scenarios where different brokers are backed by - * different Etcd clusters). Etcd is kept in sync with - * etcdserverpb.ResponseHeader. - */ +* Etcd represents the effective Etcd MVCC state under which a Gazette broker +is operating in its processing of requests and responses. Its inclusion +allows brokers to reason about relative "happened before" Revision ordering +of apparent routing conflicts in proxied or replicated requests, as well +as enabling sanity checks over equality of Etcd ClusterId (and precluding, +for example, split-brain scenarios where different brokers are backed by +different Etcd clusters). Etcd is kept in sync with +etcdserverpb.ResponseHeader. +*/ export interface HeaderEtcd { - /** - * cluster_id is the ID of the cluster. - * @format uint64 - */ - clusterId?: string; - /** - * member_id is the ID of the member. - * @format uint64 - */ - memberId?: string; - /** - * revision is the Etcd key-value store revision when the request was - * applied. - * @format int64 - */ - revision?: string; - /** - * raft_term is the raft term when the request was applied. - * @format uint64 - */ - raftTerm?: string; + /** + * cluster_id is the ID of the cluster. + * @format uint64 + */ + clusterId?: string; + /** + * member_id is the ID of the member. + * @format uint64 + */ + memberId?: string; + /** + * revision is the Etcd key-value store revision when the request was + * applied. + * @format int64 + */ + revision?: string; + /** + * raft_term is the raft term when the request was applied. + * @format uint64 + */ + raftTerm?: string; } -/** Journals of the response. */ +/** + * Journals of the response. + */ export interface ListResponseJournal { - /** JournalSpec describes a Journal and its configuration. */ - spec?: ProtocolJournalSpec; - /** - * Current ModRevision of the JournalSpec. - * @format int64 - */ - modRevision?: string; - /** Route of the journal, including endpoints. */ - route?: ProtocolRoute; + /** JournalSpec describes a Journal and its configuration. */ + spec?: ProtocolJournalSpec; + /** + * Current ModRevision of the JournalSpec. + * @format int64 + */ + modRevision?: string; + /** Route of the journal, including endpoints. */ + route?: ProtocolRoute; } -/** ID composes a zone and a suffix to uniquely identify a ProcessSpec. */ +/** + * ID composes a zone and a suffix to uniquely identify a ProcessSpec. + */ export interface ProcessSpecID { - /** - * "Zone" in which the process is running. Zones may be AWS, Azure, or - * Google Cloud Platform zone identifiers, or rack locations within a colo, - * or given some other custom meaning. Gazette will replicate across - * multiple zones, and seeks to minimize traffic which must cross zones (for - * example, by proxying reads to a broker in the current zone). - */ - zone?: string; - /** - * Unique suffix of the process within |zone|. It is permissible for a - * suffix value to repeat across zones, but never within zones. In practice, - * it's recommended to use a FQDN, Kubernetes Pod name, or comparable unique - * and self-describing value as the ID suffix. - */ - suffix?: string; + /** + * "Zone" in which the process is running. Zones may be AWS, Azure, or + * Google Cloud Platform zone identifiers, or rack locations within a colo, + * or given some other custom meaning. Gazette will replicate across + * multiple zones, and seeks to minimize traffic which must cross zones (for + * example, by proxying reads to a broker in the current zone). + */ + zone?: string; + /** + * Unique suffix of the process within |zone|. It is permissible for a + * suffix value to repeat across zones, but never within zones. In practice, + * it's recommended to use a FQDN, Kubernetes Pod name, or comparable unique + * and self-describing value as the ID suffix. + */ + suffix?: string; } export interface ProtobufAny { - typeUrl?: string; - /** @format byte */ - value?: string; + typeUrl?: string; + /** @format byte */ + value?: string; } -/** AppendResponse is the unary response message of the broker Append RPC. */ +/** + * AppendResponse is the unary response message of the broker Append RPC. + */ export interface ProtocolAppendResponse { - /** Status of the Append RPC. */ - status?: ProtocolStatus; - /** Header of the response. */ - header?: ProtocolHeader; - /** - * If status is OK, then |commit| is the Fragment which places the - * committed Append content within the Journal. - */ - commit?: ProtocolFragment; - /** Current registers of the journal. */ - registers?: ProtocolLabelSet; - /** - * Total number of RPC content chunks processed in this append. - * @format int64 - */ - totalChunks?: string; - /** - * Number of content chunks which were delayed by journal flow control. - * @format int64 - */ - delayedChunks?: string; + /** Status of the Append RPC. */ + status?: ProtocolStatus; + /** Header of the response. */ + header?: ProtocolHeader; + /** + * If status is OK, then |commit| is the Fragment which places the + * committed Append content within the Journal. + */ + commit?: ProtocolFragment; + /** Current registers of the journal. */ + registers?: ProtocolLabelSet; + /** + * Total number of RPC content chunks processed in this append. + * @format int64 + */ + totalChunks?: string; + /** + * Number of content chunks which were delayed by journal flow control. + * @format int64 + */ + delayedChunks?: string; } -/** ApplyResponse is the unary response message of the broker Apply RPC. */ +/** + * ApplyResponse is the unary response message of the broker Apply RPC. + */ export interface ProtocolApplyResponse { - /** Status of the Apply RPC. */ - status?: ProtocolStatus; - /** Header of the response. */ - header?: ProtocolHeader; + /** Status of the Apply RPC. */ + status?: ProtocolStatus; + /** Header of the response. */ + header?: ProtocolHeader; } /** - * CompressionCode defines codecs known to Gazette. - * - * - INVALID: INVALID is the zero-valued CompressionCodec, and is not a valid codec. - * - NONE: NONE encodes Fragments without any applied compression, with default suffix - * ".raw". - * - GZIP: GZIP encodes Fragments using the Gzip library, with default suffix ".gz". - * - ZSTANDARD: ZSTANDARD encodes Fragments using the ZStandard library, with default - * suffix ".zst". - * - SNAPPY: SNAPPY encodes Fragments using the Snappy library, with default suffix - * ".sz". - * - GZIP_OFFLOAD_DECOMPRESSION: GZIP_OFFLOAD_DECOMPRESSION is the GZIP codec with additional behavior - * around reads and writes to remote Fragment stores, designed to offload - * the work of decompression onto compatible stores. Specifically: - * * Fragments are written with a "Content-Encoding: gzip" header. - * * Client read requests are made with "Accept-Encoding: identity". - * This can be helpful in contexts where reader IO bandwidth to the storage - * API is unconstrained, as the cost of decompression is offloaded to the - * store and CPU-intensive batch readers may receive a parallelism benefit. - * While this codec may provide substantial read-time performance - * improvements, it is an advanced configuration and the "Content-Encoding" - * header handling can be subtle and sometimes confusing. It uses the default - * suffix ".gzod". - * @default "INVALID" - */ +* CompressionCode defines codecs known to Gazette. + + - INVALID: INVALID is the zero-valued CompressionCodec, and is not a valid codec. + - NONE: NONE encodes Fragments without any applied compression, with default suffix +".raw". + - GZIP: GZIP encodes Fragments using the Gzip library, with default suffix ".gz". + - ZSTANDARD: ZSTANDARD encodes Fragments using the ZStandard library, with default +suffix ".zst". + - SNAPPY: SNAPPY encodes Fragments using the Snappy library, with default suffix +".sz". + - GZIP_OFFLOAD_DECOMPRESSION: GZIP_OFFLOAD_DECOMPRESSION is the GZIP codec with additional behavior +around reads and writes to remote Fragment stores, designed to offload +the work of decompression onto compatible stores. Specifically: + * Fragments are written with a "Content-Encoding: gzip" header. + * Client read requests are made with "Accept-Encoding: identity". +This can be helpful in contexts where reader IO bandwidth to the storage +API is unconstrained, as the cost of decompression is offloaded to the +store and CPU-intensive batch readers may receive a parallelism benefit. +While this codec may provide substantial read-time performance +improvements, it is an advanced configuration and the "Content-Encoding" +header handling can be subtle and sometimes confusing. It uses the default +suffix ".gzod". +*/ export type ProtocolCompressionCodec = - | "INVALID" - | "NONE" - | "GZIP" - | "ZSTANDARD" - | "SNAPPY" - | "GZIP_OFFLOAD_DECOMPRESSION"; + | 'INVALID' + | 'NONE' + | 'GZIP' + | 'ZSTANDARD' + | 'SNAPPY' + | 'GZIP_OFFLOAD_DECOMPRESSION'; /** - * Fragment is a content-addressed description of a contiguous Journal span, - * defined by the [begin, end) offset range covered by the Fragment and the - * SHA1 sum of the corresponding Journal content. - */ +* Fragment is a content-addressed description of a contiguous Journal span, +defined by the [begin, end) offset range covered by the Fragment and the +SHA1 sum of the corresponding Journal content. +*/ export interface ProtocolFragment { - /** Journal of the Fragment. */ - journal?: string; - /** - * Begin (inclusive) and end (exclusive) offset of the Fragment within the - * Journal. - * @format int64 - */ - begin?: string; - /** @format int64 */ - end?: string; - /** SHA1 sum of the Fragment's content. */ - sum?: ProtocolSHA1Sum; - /** Codec with which the Fragment's content is compressed. */ - compressionCodec?: ProtocolCompressionCodec; - /** - * Fragment store which backs the Fragment. Empty if the Fragment has yet to - * be persisted and is still local to a Broker. - */ - backingStore?: string; - /** - * Modification timestamp of the Fragment within the backing store, - * represented as seconds since the epoch. - * @format int64 - */ - modTime?: string; - /** - * Path postfix under which the fragment is persisted to the store. - * The complete Fragment store path is built from any path components of the - * backing store, followed by the journal name, followed by the path postfix. - */ - pathPostfix?: string; + /** Journal of the Fragment. */ + journal?: string; + /** + * Begin (inclusive) and end (exclusive) offset of the Fragment within the + * Journal. + * @format int64 + */ + begin?: string; + /** @format int64 */ + end?: string; + /** SHA1 sum of the Fragment's content. */ + sum?: ProtocolSHA1Sum; + /** Codec with which the Fragment's content is compressed. */ + compressionCodec?: ProtocolCompressionCodec; + /** + * Fragment store which backs the Fragment. Empty if the Fragment has yet to + * be persisted and is still local to a Broker. + */ + backingStore?: string; + /** + * Modification timestamp of the Fragment within the backing store, + * represented as seconds since the epoch. + * @format int64 + */ + modTime?: string; + /** + * Path postfix under which the fragment is persisted to the store. + * The complete Fragment store path is built from any path components of the + * backing store, followed by the journal name, followed by the path postfix. + */ + pathPostfix?: string; } /** - * FragmentsResponse is the unary response message of the broker ListFragments - * RPC. - */ +* FragmentsResponse is the unary response message of the broker ListFragments +RPC. +*/ export interface ProtocolFragmentsResponse { - /** Status of the Apply RPC. */ - status?: ProtocolStatus; - /** Header of the response. */ - header?: ProtocolHeader; - fragments?: FragmentsResponseFragment[]; - /** - * The NextPageToke value to be returned on subsequent Fragments requests. If - * the value is zero then there are no more fragments to be returned for this - * page. - * @format int64 - */ - nextPageToken?: string; + /** Status of the Apply RPC. */ + status?: ProtocolStatus; + /** Header of the response. */ + header?: ProtocolHeader; + fragments?: FragmentsResponseFragment[]; + /** + * The NextPageToke value to be returned on subsequent Fragments requests. If + * the value is zero then there are no more fragments to be returned for this + * page. + * @format int64 + */ + nextPageToken?: string; } /** - * Header captures metadata such as the process responsible for processing - * an RPC, and its effective Etcd state. - */ +* Header captures metadata such as the process responsible for processing +an RPC, and its effective Etcd state. +*/ export interface ProtocolHeader { - /** - * ID of the process responsible for request processing. May be empty iff - * Header is being used within a proxied request, and that request may be - * dispatched to any member of the Route. - */ - processId?: ProcessSpecID; - /** - * Route of processes specifically responsible for this RPC, or an empty Route - * if any process is capable of serving the RPC. - */ - route?: ProtocolRoute; - /** - * Etcd represents the effective Etcd MVCC state under which a Gazette broker - * is operating in its processing of requests and responses. Its inclusion - * allows brokers to reason about relative "happened before" Revision ordering - * of apparent routing conflicts in proxied or replicated requests, as well - * as enabling sanity checks over equality of Etcd ClusterId (and precluding, - * for example, split-brain scenarios where different brokers are backed by - * different Etcd clusters). Etcd is kept in sync with - * etcdserverpb.ResponseHeader. - */ - etcd?: HeaderEtcd; + /** + * ID of the process responsible for request processing. May be empty iff + * Header is being used within a proxied request, and that request may be + * dispatched to any member of the Route. + */ + processId?: ProcessSpecID; + /** + * Route of processes specifically responsible for this RPC, or an empty Route + * if any process is capable of serving the RPC. + */ + route?: ProtocolRoute; + /** + * Etcd represents the effective Etcd MVCC state under which a Gazette broker + * is operating in its processing of requests and responses. Its inclusion + * allows brokers to reason about relative "happened before" Revision ordering + * of apparent routing conflicts in proxied or replicated requests, as well + * as enabling sanity checks over equality of Etcd ClusterId (and precluding, + * for example, split-brain scenarios where different brokers are backed by + * different Etcd clusters). Etcd is kept in sync with + * etcdserverpb.ResponseHeader. + */ + etcd?: HeaderEtcd; } -/** JournalSpec describes a Journal and its configuration. */ +/** + * JournalSpec describes a Journal and its configuration. + */ export interface ProtocolJournalSpec { - /** Name of the Journal. */ - name?: string; - /** - * Desired replication of this Journal. This defines the Journal's tolerance - * to broker failures before data loss can occur (eg, a replication factor - * of three means two failures are tolerated). - * @format int32 - */ - replication?: number; - /** - * User-defined Labels of this JournalSpec. Two label names are reserved - * and may not be used within a JournalSpec's Labels: "name" and "prefix". - */ - labels?: ProtocolLabelSet; - /** - * Fragment is JournalSpec configuration which pertains to the creation, - * persistence, and indexing of the Journal's Fragments. - */ - fragment?: ProtocolJournalSpecFragment; - /** - * Flags of the Journal, as a combination of Flag enum values. The Flag enum - * is not used directly, as protobuf enums do not allow for or'ed bitfields. - * @format int64 - */ - flags?: number; - /** - * Maximum rate, in bytes-per-second, at which appends of this journal will - * be processed. If zero (the default), no rate limiting is applied. A global - * rate limit still may be in effect, in which case the effective rate is the - * smaller of the journal vs global rate. - * @format int64 - */ - maxAppendRate?: string; + /** Name of the Journal. */ + name?: string; + /** + * Desired replication of this Journal. This defines the Journal's tolerance + * to broker failures before data loss can occur (eg, a replication factor + * of three means two failures are tolerated). + * @format int32 + */ + replication?: number; + /** + * User-defined Labels of this JournalSpec. Two label names are reserved + * and may not be used within a JournalSpec's Labels: "name" and "prefix". + */ + labels?: ProtocolLabelSet; + /** + * Fragment is JournalSpec configuration which pertains to the creation, + * persistence, and indexing of the Journal's Fragments. + */ + fragment?: ProtocolJournalSpecFragment; + /** + * Flags of the Journal, as a combination of Flag enum values. The Flag enum + * is not used directly, as protobuf enums do not allow for or'ed bitfields. + * @format int64 + */ + flags?: number; + /** + * Maximum rate, in bytes-per-second, at which appends of this journal will + * be processed. If zero (the default), no rate limiting is applied. A global + * rate limit still may be in effect, in which case the effective rate is the + * smaller of the journal vs global rate. + * @format int64 + */ + maxAppendRate?: string; } /** - * Fragment is JournalSpec configuration which pertains to the creation, - * persistence, and indexing of the Journal's Fragments. - */ +* Fragment is JournalSpec configuration which pertains to the creation, +persistence, and indexing of the Journal's Fragments. +*/ export interface ProtocolJournalSpecFragment { - /** - * Target content length of each Fragment. In normal operation after - * Fragments reach at least this length, they will be closed and new ones - * begun. Note lengths may be smaller at times (eg, due to changes in - * Journal routing topology). Content length differs from Fragment file - * size, in that the former reflects uncompressed bytes. - * @format int64 - */ - length?: string; - /** Codec used to compress Journal Fragments. */ - compressionCodec?: ProtocolCompressionCodec; - /** - * Storage backend base path for this Journal's Fragments. Must be in URL - * form, with the choice of backend defined by the scheme. The full path of - * a Journal's Fragment is derived by joining the store path with the - * Fragment's ContentPath. Eg, given a fragment_store of - * "s3://My-AWS-bucket/a/prefix" and a JournalSpec of name "my/journal", - * a complete Fragment path might be: - * "s3://My-AWS-bucket/a/prefix/my/journal/000123-000456-789abcdef.gzip - * Multiple stores may be specified, in which case the Journal's Fragments - * are the union of all Fragments present across all stores, and new - * Fragments always persist to the first specified store. This can be - * helpful in performing incremental migrations, where new Journal content - * is written to the new store, while content in the old store remains - * available (and, depending on fragment_retention or recovery log pruning, - * may eventually be removed). - * - * If no stores are specified, the Journal is still use-able but will - * not persist Fragments to any a backing fragment store. This allows for - * real-time streaming use cases where reads of historical data are not - * needed. - */ - stores?: string[]; - /** - * Interval of time between refreshes of remote Fragment listings from - * configured fragment_stores. - */ - refreshInterval?: string; - /** - * Retention duration for historical Fragments of this Journal within the - * Fragment stores. If less than or equal to zero, Fragments are retained - * indefinitely. - */ - retention?: string; - /** - * Flush interval defines a uniform UTC time segment which, when passed, - * will prompt brokers to close and persist a fragment presently being - * written. - * - * Flush interval may be helpful in integrating the journal with a regularly - * scheduled batch work-flow which processes new files from the fragment - * store and has no particular awareness of Gazette. For example, setting - * flush_interval to 3600s will cause brokers to persist their present - * fragment on the hour, every hour, even if it has not yet reached its - * target length. A batch work-flow running at 5 minutes past the hour is - * then reasonably assured of seeing all events from the past hour. - * - * See also "gazctl journals fragments --help" for more discussion. - */ - flushInterval?: string; - /** - * Path postfix template is a Go template which evaluates to a partial - * path under which fragments are persisted to the store. A complete - * fragment path is constructed by appending path components from the - * fragment store, then the journal name, and then the postfix template. - * Path post-fixes can help in maintaining Hive compatible partitioning - * over fragment creation time. The fields ".Spool" and ".JournalSpec" - * are available for introspection in the template. For example, - * to partition on the UTC date and hour of creation, use: - * date={{ .Spool.FirstAppendTime.Format "2006-01-02" }}/hour={{ - * .Spool.FirstAppendTime.Format "15" }} - * - * Which will produce a path postfix like "date=2019-11-19/hour=22". - */ - pathPostfixTemplate?: string; + /** + * Target content length of each Fragment. In normal operation after + * Fragments reach at least this length, they will be closed and new ones + * begun. Note lengths may be smaller at times (eg, due to changes in + * Journal routing topology). Content length differs from Fragment file + * size, in that the former reflects uncompressed bytes. + * @format int64 + */ + length?: string; + /** Codec used to compress Journal Fragments. */ + compressionCodec?: ProtocolCompressionCodec; + /** + * Storage backend base path for this Journal's Fragments. Must be in URL + * form, with the choice of backend defined by the scheme. The full path of + * a Journal's Fragment is derived by joining the store path with the + * Fragment's ContentPath. Eg, given a fragment_store of + * "s3://My-AWS-bucket/a/prefix" and a JournalSpec of name "my/journal", + * a complete Fragment path might be: + * "s3://My-AWS-bucket/a/prefix/my/journal/000123-000456-789abcdef.gzip + * Multiple stores may be specified, in which case the Journal's Fragments + * are the union of all Fragments present across all stores, and new + * Fragments always persist to the first specified store. This can be + * helpful in performing incremental migrations, where new Journal content + * is written to the new store, while content in the old store remains + * available (and, depending on fragment_retention or recovery log pruning, + * may eventually be removed). + * + * If no stores are specified, the Journal is still use-able but will + * not persist Fragments to any a backing fragment store. This allows for + * real-time streaming use cases where reads of historical data are not + * needed. + */ + stores?: string[]; + /** + * Interval of time between refreshes of remote Fragment listings from + * configured fragment_stores. + */ + refreshInterval?: string; + /** + * Retention duration for historical Fragments of this Journal within the + * Fragment stores. If less than or equal to zero, Fragments are retained + * indefinitely. + */ + retention?: string; + /** + * Flush interval defines a uniform UTC time segment which, when passed, + * will prompt brokers to close and persist a fragment presently being + * written. + * + * Flush interval may be helpful in integrating the journal with a regularly + * scheduled batch work-flow which processes new files from the fragment + * store and has no particular awareness of Gazette. For example, setting + * flush_interval to 3600s will cause brokers to persist their present + * fragment on the hour, every hour, even if it has not yet reached its + * target length. A batch work-flow running at 5 minutes past the hour is + * then reasonably assured of seeing all events from the past hour. + * See also "gazctl journals fragments --help" for more discussion. + */ + flushInterval?: string; + /** + * Path postfix template is a Go template which evaluates to a partial + * path under which fragments are persisted to the store. A complete + * fragment path is constructed by appending path components from the + * fragment store, then the journal name, and then the postfix template. + * Path post-fixes can help in maintaining Hive compatible partitioning + * over fragment creation time. The fields ".Spool" and ".JournalSpec" + * are available for introspection in the template. For example, + * to partition on the UTC date and hour of creation, use: + * date={{ .Spool.FirstAppendTime.Format "2006-01-02" }}/hour={{ + * .Spool.FirstAppendTime.Format "15" }} + * + * Which will produce a path postfix like "date=2019-11-19/hour=22". + */ + pathPostfixTemplate?: string; } /** - * Label defines a key & value pair which can be attached to entities like - * JournalSpecs and BrokerSpecs. Labels may be used to provide identifying - * attributes which do not directly imply semantics to the core system, but - * are meaningful to users or for higher-level Gazette tools. - */ +* Label defines a key & value pair which can be attached to entities like +JournalSpecs and BrokerSpecs. Labels may be used to provide identifying +attributes which do not directly imply semantics to the core system, but +are meaningful to users or for higher-level Gazette tools. +*/ export interface ProtocolLabel { - name?: string; - value?: string; + name?: string; + value?: string; } -/** LabelSelector defines a filter over LabelSets. */ +/** + * LabelSelector defines a filter over LabelSets. + */ export interface ProtocolLabelSelector { - /** - * Include is Labels which must be matched for a LabelSet to be selected. If - * empty, all Labels are included. An include Label with empty ("") value is - * matched by a Label of the same name having any value. - */ - include?: ProtocolLabelSet; - /** - * Exclude is Labels which cannot be matched for a LabelSet to be selected. If - * empty, no Labels are excluded. An exclude Label with empty ("") value - * excludes a Label of the same name having any value. - */ - exclude?: ProtocolLabelSet; + /** + * Include is Labels which must be matched for a LabelSet to be selected. If + * empty, all Labels are included. An include Label with empty ("") value is + * matched by a Label of the same name having any value. + */ + include?: ProtocolLabelSet; + /** + * Exclude is Labels which cannot be matched for a LabelSet to be selected. If + * empty, no Labels are excluded. An exclude Label with empty ("") value + * excludes a Label of the same name having any value. + */ + exclude?: ProtocolLabelSet; } -/** LabelSet is a collection of labels and their values. */ +/** + * LabelSet is a collection of labels and their values. + */ export interface ProtocolLabelSet { - /** Labels of the set. Instances must be unique and sorted over (Name, Value). */ - labels?: ProtocolLabel[]; + /** Labels of the set. Instances must be unique and sorted over (Name, Value). */ + labels?: ProtocolLabel[]; } -/** ListRequest is the unary request message of the broker List RPC. */ +/** + * ListRequest is the unary request message of the broker List RPC. + */ export interface ProtocolListRequest { - /** - * Selector optionally refines the set of journals which will be enumerated. - * If zero-valued, all journals are returned. Otherwise, only JournalSpecs - * matching the LabelSelector will be returned. Two meta-labels "name" and - * "prefix" are additionally supported by the selector, where: - * * name=examples/a-name will match a JournalSpec with Name - * "examples/a-name" - * * prefix=examples/ will match any JournalSpec having prefix "examples/". - * The prefix Label value must end in '/'. - */ - selector?: ProtocolLabelSelector; + /** + * Selector optionally refines the set of journals which will be enumerated. + * If zero-valued, all journals are returned. Otherwise, only JournalSpecs + * matching the LabelSelector will be returned. Two meta-labels "name" and + * "prefix" are additionally supported by the selector, where: + * * name=examples/a-name will match a JournalSpec with Name + * "examples/a-name" + * * prefix=examples/ will match any JournalSpec having prefix "examples/". + * The prefix Label value must end in '/'. + */ + selector?: ProtocolLabelSelector; } -/** ListResponse is the unary response message of the broker List RPC. */ +/** + * ListResponse is the unary response message of the broker List RPC. + */ export interface ProtocolListResponse { - /** Status of the List RPC. */ - status?: ProtocolStatus; - /** Header of the response. */ - header?: ProtocolHeader; - journals?: ListResponseJournal[]; + /** Status of the List RPC. */ + status?: ProtocolStatus; + /** Header of the response. */ + header?: ProtocolHeader; + journals?: ListResponseJournal[]; } -/** ReadRequest is the unary request message of the broker Read RPC. */ +/** + * ReadRequest is the unary request message of the broker Read RPC. + */ export interface ProtocolReadRequest { - /** Header is attached by a proxying broker peer. */ - header?: ProtocolHeader; - /** Journal to be read. */ - journal?: string; - /** - * Desired offset to begin reading from. Value -1 has special handling, where - * the read is performed from the current write head. All other positive - * values specify a desired exact byte offset to read from. If the offset is - * not available (eg, because it represents a portion of Journal which has - * been permanently deleted), the broker will return the next available - * offset. Callers should therefore always inspect the ReadResponse offset. - * @format int64 - */ - offset?: string; - /** - * Whether the operation should block until content becomes available. - * OFFSET_NOT_YET_AVAILABLE is returned if a non-blocking read has no ready - * content. - */ - block?: boolean; - /** - * If do_not_proxy is true, the broker will not proxy the read to another - * broker, or open and proxy a remote Fragment on the client's behalf. - */ - doNotProxy?: boolean; - /** - * If metadata_only is true, the broker will respond with Journal and - * Fragment metadata but not content. - */ - metadataOnly?: boolean; - /** - * Offset to read through. If zero, then the read end offset is unconstrained. - * @format int64 - */ - endOffset?: string; + /** Header is attached by a proxying broker peer. */ + header?: ProtocolHeader; + /** Journal to be read. */ + journal?: string; + /** + * Desired offset to begin reading from. Value -1 has special handling, where + * the read is performed from the current write head. All other positive + * values specify a desired exact byte offset to read from. If the offset is + * not available (eg, because it represents a portion of Journal which has + * been permanently deleted), the broker will return the next available + * offset. Callers should therefore always inspect the ReadResponse offset. + * @format int64 + */ + offset?: string; + /** + * Whether the operation should block until content becomes available. + * OFFSET_NOT_YET_AVAILABLE is returned if a non-blocking read has no ready + * content. + */ + block?: boolean; + /** + * If do_not_proxy is true, the broker will not proxy the read to another + * broker, or open and proxy a remote Fragment on the client's behalf. + */ + doNotProxy?: boolean; + /** + * If metadata_only is true, the broker will respond with Journal and + * Fragment metadata but not content. + */ + metadataOnly?: boolean; + /** + * Offset to read through. If zero, then the read end offset is unconstrained. + * @format int64 + */ + endOffset?: string; + /** + * BeginModTime is an optional inclusive lower bound on the modification + * timestamps of fragments read from the backing store, represented as + * seconds since the epoch. The request Offset will be advanced as-needed + * to skip persisted Fragments having a modication time before the bound. + * @format int64 + */ + beginModTime?: string; } /** - * ReadResponse is the streamed response message of the broker Read RPC. - * Responses messages are of two types: - * * "Metadata" messages, which conveys the journal Fragment addressed by the - * request which is ready to be read. - * * "Chunk" messages, which carry associated journal Fragment content bytes. - * - * A metadata message specifying a Fragment always precedes all "chunks" of the - * Fragment's content. Response streams may be very long lived, having many - * metadata and accompanying chunk messages. The reader may also block for long - * periods of time awaiting the next metadata message (eg, if the next offset - * hasn't yet committed). However once a metadata message is read, the reader - * is assured that its associated chunk messages are immediately forthcoming. - */ +* ReadResponse is the streamed response message of the broker Read RPC. +Responses messages are of two types: +* * "Metadata" messages, which conveys the journal Fragment addressed by the + request which is ready to be read. +* "Chunk" messages, which carry associated journal Fragment content bytes. + +A metadata message specifying a Fragment always precedes all "chunks" of the +Fragment's content. Response streams may be very long lived, having many +metadata and accompanying chunk messages. The reader may also block for long +periods of time awaiting the next metadata message (eg, if the next offset +hasn't yet committed). However once a metadata message is read, the reader +is assured that its associated chunk messages are immediately forthcoming. +*/ export interface ProtocolReadResponse { - /** Status of the Read RPC. */ - status?: ProtocolStatus; - /** - * Header of the response. Accompanies the first ReadResponse of the response - * stream. - */ - header?: ProtocolHeader; - /** - * The effective offset of the read. See ReadRequest offset. - * @format int64 - */ - offset?: string; - /** - * The offset to next be written, by the next append transaction served by - * broker. In other words, the last offset through which content is - * available to be read from the Journal. This is a metadata field and will - * not be returned with a content response. - * @format int64 - */ - writeHead?: string; - /** - * Fragment to which the offset was mapped. This is a metadata field and will - * not be returned with a content response. - */ - fragment?: ProtocolFragment; - /** If Fragment is remote, a URL from which it may be directly read. */ - fragmentUrl?: string; - /** - * Content chunks of the read. - * @format byte - */ - content?: string; + /** Status of the Read RPC. */ + status?: ProtocolStatus; + /** + * Header of the response. Accompanies the first ReadResponse of the response + * stream. + */ + header?: ProtocolHeader; + /** + * The effective offset of the read. See ReadRequest offset. + * @format int64 + */ + offset?: string; + /** + * The offset to next be written, by the next append transaction served by + * broker. In other words, the last offset through which content is + * available to be read from the Journal. This is a metadata field and will + * not be returned with a content response. + * @format int64 + */ + writeHead?: string; + /** + * Fragment to which the offset was mapped. This is a metadata field and will + * not be returned with a content response. + */ + fragment?: ProtocolFragment; + /** If Fragment is remote, a URL from which it may be directly read. */ + fragmentUrl?: string; + /** + * Content chunks of the read. + * @format byte + */ + content?: string; } /** - * ReplicateResponse is the streamed response message of the broker's internal - * Replicate RPC. Each message is a 1:1 response to a previously read "proposal" - * ReplicateRequest with |acknowledge| set. - */ +* ReplicateResponse is the streamed response message of the broker's internal +Replicate RPC. Each message is a 1:1 response to a previously read "proposal" +ReplicateRequest with |acknowledge| set. +*/ export interface ProtocolReplicateResponse { - /** Status of the Replicate RPC. */ - status?: ProtocolStatus; - /** - * Header of the response. Accompanies the first ReplicateResponse of the - * response stream. - */ - header?: ProtocolHeader; - /** - * If status is PROPOSAL_MISMATCH, then |fragment| is the replica's current - * journal Fragment, and either it or |registers| will differ from the - * primary's proposal. - */ - fragment?: ProtocolFragment; - /** - * If status is PROPOSAL_MISMATCH, then |registers| are the replica's current - * journal registers. - */ - registers?: ProtocolLabelSet; + /** Status of the Replicate RPC. */ + status?: ProtocolStatus; + /** + * Header of the response. Accompanies the first ReplicateResponse of the + * response stream. + */ + header?: ProtocolHeader; + /** + * If status is PROPOSAL_MISMATCH, then |fragment| is the replica's current + * journal Fragment, and either it or |registers| will differ from the + * primary's proposal. + */ + fragment?: ProtocolFragment; + /** + * If status is PROPOSAL_MISMATCH, then |registers| are the replica's current + * journal registers. + */ + registers?: ProtocolLabelSet; } -/** Route captures the current topology of an item and the processes serving it. */ +/** + * Route captures the current topology of an item and the processes serving it. + */ export interface ProtocolRoute { - /** Members of the Route, ordered on ascending ProcessSpec.ID (zone, suffix). */ - members?: ProcessSpecID[]; - /** - * Index of the ProcessSpec serving as primary within |members|, - * or -1 of no member is currently primary. - * @format int32 - */ - primary?: number; - /** - * Endpoints of each Route member. If not empty, |endpoints| has the same - * length and order as |members|, and captures the endpoint of each one. - */ - endpoints?: string[]; + /** Members of the Route, ordered on ascending ProcessSpec.ID (zone, suffix). */ + members?: ProcessSpecID[]; + /** + * Index of the ProcessSpec serving as primary within |members|, + * or -1 of no member is currently primary. + * @format int32 + */ + primary?: number; + /** + * Endpoints of each Route member. If not empty, |endpoints| has the same + * length and order as |members|, and captures the endpoint of each one. + */ + endpoints?: string[]; } -/** SHA1Sum is a 160-bit SHA1 digest. */ +/** + * SHA1Sum is a 160-bit SHA1 digest. + */ export interface ProtocolSHA1Sum { - /** @format uint64 */ - part1?: string; - /** @format uint64 */ - part2?: string; - /** @format int64 */ - part3?: number; + /** @format uint64 */ + part1?: string; + /** @format uint64 */ + part2?: string; + /** @format int64 */ + part3?: number; } /** - * Status is a response status code, used universally across Gazette RPC APIs. - * - * - JOURNAL_NOT_FOUND: The named journal does not exist. - * - NO_JOURNAL_PRIMARY_BROKER: There is no current primary broker for the journal. This is a temporary - * condition which should quickly resolve, assuming sufficient broker - * capacity. - * - NOT_JOURNAL_PRIMARY_BROKER: The present broker is not the assigned primary broker for the journal. - * - NOT_JOURNAL_BROKER: The present broker is not an assigned broker for the journal. - * - INSUFFICIENT_JOURNAL_BROKERS: There are an insufficient number of assigned brokers for the journal - * to meet its required replication. - * - OFFSET_NOT_YET_AVAILABLE: The requested offset is not yet available. This indicates either that the - * offset has not yet been written, or that the broker is not yet aware of a - * written fragment covering the offset. Returned only by non-blocking reads. - * - WRONG_ROUTE: The peer disagrees with the Route accompanying a ReplicateRequest. - * - PROPOSAL_MISMATCH: The peer disagrees with the proposal accompanying a ReplicateRequest. - * - ETCD_TRANSACTION_FAILED: The Etcd transaction failed. Returned by Update RPC when an - * expect_mod_revision of the UpdateRequest differs from the current - * ModRevision of the JournalSpec within the store. - * - NOT_ALLOWED: A disallowed journal access was attempted (eg, a write where the - * journal disables writes, or read where journals disable reads). - * - WRONG_APPEND_OFFSET: The Append is refused because its requested offset is not equal - * to the furthest written offset of the journal. - * - INDEX_HAS_GREATER_OFFSET: The Append is refused because the replication pipeline tracks a smaller - * journal offset than that of the remote fragment index. This indicates - * that journal replication consistency has been lost in the past, due to - * too many broker or Etcd failures. - * - REGISTER_MISMATCH: The Append is refused because a registers selector was provided with the - * request, but it was not matched by current register values of the journal. - * @default "OK" - */ +* Status is a response status code, used universally across Gazette RPC APIs. + + - JOURNAL_NOT_FOUND: The named journal does not exist. + - NO_JOURNAL_PRIMARY_BROKER: There is no current primary broker for the journal. This is a temporary +condition which should quickly resolve, assuming sufficient broker +capacity. + - NOT_JOURNAL_PRIMARY_BROKER: The present broker is not the assigned primary broker for the journal. + - NOT_JOURNAL_BROKER: The present broker is not an assigned broker for the journal. + - INSUFFICIENT_JOURNAL_BROKERS: There are an insufficient number of assigned brokers for the journal +to meet its required replication. + - OFFSET_NOT_YET_AVAILABLE: The requested offset is not yet available. This indicates either that the +offset has not yet been written, or that the broker is not yet aware of a +written fragment covering the offset. Returned only by non-blocking reads. + - WRONG_ROUTE: The peer disagrees with the Route accompanying a ReplicateRequest. + - PROPOSAL_MISMATCH: The peer disagrees with the proposal accompanying a ReplicateRequest. + - ETCD_TRANSACTION_FAILED: The Etcd transaction failed. Returned by Update RPC when an +expect_mod_revision of the UpdateRequest differs from the current +ModRevision of the JournalSpec within the store. + - NOT_ALLOWED: A disallowed journal access was attempted (eg, a write where the +journal disables writes, or read where journals disable reads). + - WRONG_APPEND_OFFSET: The Append is refused because its requested offset is not equal +to the furthest written offset of the journal. + - INDEX_HAS_GREATER_OFFSET: The Append is refused because the replication pipeline tracks a smaller +journal offset than that of the remote fragment index. This indicates +that journal replication consistency has been lost in the past, due to +too many broker or Etcd failures. + - REGISTER_MISMATCH: The Append is refused because a registers selector was provided with the +request, but it was not matched by current register values of the journal. +*/ export type ProtocolStatus = - | "OK" - | "JOURNAL_NOT_FOUND" - | "NO_JOURNAL_PRIMARY_BROKER" - | "NOT_JOURNAL_PRIMARY_BROKER" - | "NOT_JOURNAL_BROKER" - | "INSUFFICIENT_JOURNAL_BROKERS" - | "OFFSET_NOT_YET_AVAILABLE" - | "WRONG_ROUTE" - | "PROPOSAL_MISMATCH" - | "ETCD_TRANSACTION_FAILED" - | "NOT_ALLOWED" - | "WRONG_APPEND_OFFSET" - | "INDEX_HAS_GREATER_OFFSET" - | "REGISTER_MISMATCH"; + | 'OK' + | 'JOURNAL_NOT_FOUND' + | 'NO_JOURNAL_PRIMARY_BROKER' + | 'NOT_JOURNAL_PRIMARY_BROKER' + | 'NOT_JOURNAL_BROKER' + | 'INSUFFICIENT_JOURNAL_BROKERS' + | 'OFFSET_NOT_YET_AVAILABLE' + | 'WRONG_ROUTE' + | 'PROPOSAL_MISMATCH' + | 'ETCD_TRANSACTION_FAILED' + | 'NOT_ALLOWED' + | 'WRONG_APPEND_OFFSET' + | 'INDEX_HAS_GREATER_OFFSET' + | 'REGISTER_MISMATCH'; export interface RuntimeError { - error?: string; - /** @format int32 */ - code?: number; - message?: string; - details?: ProtobufAny[]; + error?: string; + /** @format int32 */ + code?: number; + message?: string; + details?: ProtobufAny[]; } export interface RuntimeStreamError { - /** @format int32 */ - grpcCode?: number; - /** @format int32 */ - httpCode?: number; - message?: string; - httpStatus?: string; - details?: ProtobufAny[]; + /** @format int32 */ + grpcCode?: number; + /** @format int32 */ + httpCode?: number; + message?: string; + httpStatus?: string; + details?: ProtobufAny[]; } diff --git a/client/src/gen/consumer/protocol/consumer.ts b/client/src/gen/consumer/protocol/consumer.ts index e39de6f..e8c797e 100644 --- a/client/src/gen/consumer/protocol/consumer.ts +++ b/client/src/gen/consumer/protocol/consumer.ts @@ -10,663 +10,669 @@ */ export interface GetHintsResponseResponseHints { - /** If the hints value does not exist Hints will be nil. */ - hints?: RecoverylogFSMHints; + /** If the hints value does not exist Hints will be nil. */ + hints?: RecoverylogFSMHints; } /** - * Etcd represents the effective Etcd MVCC state under which a Gazette broker - * is operating in its processing of requests and responses. Its inclusion - * allows brokers to reason about relative "happened before" Revision ordering - * of apparent routing conflicts in proxied or replicated requests, as well - * as enabling sanity checks over equality of Etcd ClusterId (and precluding, - * for example, split-brain scenarios where different brokers are backed by - * different Etcd clusters). Etcd is kept in sync with - * etcdserverpb.ResponseHeader. - */ +* Etcd represents the effective Etcd MVCC state under which a Gazette broker +is operating in its processing of requests and responses. Its inclusion +allows brokers to reason about relative "happened before" Revision ordering +of apparent routing conflicts in proxied or replicated requests, as well +as enabling sanity checks over equality of Etcd ClusterId (and precluding, +for example, split-brain scenarios where different brokers are backed by +different Etcd clusters). Etcd is kept in sync with +etcdserverpb.ResponseHeader. +*/ export interface HeaderEtcd { - /** - * cluster_id is the ID of the cluster. - * @format uint64 - */ - clusterId?: string; - /** - * member_id is the ID of the member. - * @format uint64 - */ - memberId?: string; - /** - * revision is the Etcd key-value store revision when the request was - * applied. - * @format int64 - */ - revision?: string; - /** - * raft_term is the raft term when the request was applied. - * @format uint64 - */ - raftTerm?: string; + /** + * cluster_id is the ID of the cluster. + * @format uint64 + */ + clusterId?: string; + /** + * member_id is the ID of the member. + * @format uint64 + */ + memberId?: string; + /** + * revision is the Etcd key-value store revision when the request was + * applied. + * @format int64 + */ + revision?: string; + /** + * raft_term is the raft term when the request was applied. + * @format uint64 + */ + raftTerm?: string; } -/** Shards of the response. */ +/** + * Shards of the response. + */ export interface ListResponseShard { - /** - * ShardSpec describes a shard and its configuration, and is the long-lived unit - * of work and scaling for a consumer application. Each shard is allocated to a - * one "primary" at-a-time selected from the current processes of a consumer - * application, and is re-assigned on process fault or exit. - * - * ShardSpecs describe all configuration of the shard and its processing, - * including journals to consume, configuration for processing transactions, its - * recovery log, hot standbys, etc. ShardSpecs may be further extended with - * domain-specific labels & values to further define application behavior. - * ShardSpec is-a allocator.ItemValue. - */ - spec?: ConsumerShardSpec; - /** - * Current ModRevision of the ShardSpec. - * @format int64 - */ - modRevision?: string; - /** Route of the shard, including endpoints. */ - route?: ProtocolRoute; - /** Status of each replica. Cardinality and ordering matches |route|. */ - status?: ConsumerReplicaStatus[]; + /** + * ShardSpec describes a shard and its configuration, and is the long-lived unit + * of work and scaling for a consumer application. Each shard is allocated to a + * one "primary" at-a-time selected from the current processes of a consumer + * application, and is re-assigned on process fault or exit. + * + * ShardSpecs describe all configuration of the shard and its processing, + * including journals to consume, configuration for processing transactions, its + * recovery log, hot standbys, etc. ShardSpecs may be further extended with + * domain-specific labels & values to further define application behavior. + * ShardSpec is-a allocator.ItemValue. + */ + spec?: ConsumerShardSpec; + /** + * Current ModRevision of the ShardSpec. + * @format int64 + */ + modRevision?: string; + /** Route of the shard, including endpoints. */ + route?: ProtocolRoute; + /** Status of each replica. Cardinality and ordering matches |route|. */ + status?: ConsumerReplicaStatus[]; } -/** ID composes a zone and a suffix to uniquely identify a ProcessSpec. */ +/** + * ID composes a zone and a suffix to uniquely identify a ProcessSpec. + */ export interface ProcessSpecID { - /** - * "Zone" in which the process is running. Zones may be AWS, Azure, or - * Google Cloud Platform zone identifiers, or rack locations within a colo, - * or given some other custom meaning. Gazette will replicate across - * multiple zones, and seeks to minimize traffic which must cross zones (for - * example, by proxying reads to a broker in the current zone). - */ - zone?: string; - /** - * Unique suffix of the process within |zone|. It is permissible for a - * suffix value to repeat across zones, but never within zones. In practice, - * it's recommended to use a FQDN, Kubernetes Pod name, or comparable unique - * and self-describing value as the ID suffix. - */ - suffix?: string; + /** + * "Zone" in which the process is running. Zones may be AWS, Azure, or + * Google Cloud Platform zone identifiers, or rack locations within a colo, + * or given some other custom meaning. Gazette will replicate across + * multiple zones, and seeks to minimize traffic which must cross zones (for + * example, by proxying reads to a broker in the current zone). + */ + zone?: string; + /** + * Unique suffix of the process within |zone|. It is permissible for a + * suffix value to repeat across zones, but never within zones. In practice, + * it's recommended to use a FQDN, Kubernetes Pod name, or comparable unique + * and self-describing value as the ID suffix. + */ + suffix?: string; } /** - * - BACKFILL: The replica is actively playing the historical recovery log. - * - STANDBY: The replica has finished playing the historical recovery log and is - * live-tailing it to locally mirror recorded operations as they are - * produced. It can take over as primary at any time. - * - * Shards not having recovery logs immediately transition to STANDBY. - * - PRIMARY: The replica is actively serving as primary. - * - FAILED: The replica has encountered an unrecoverable error. - * @default "IDLE" - */ -export type ReplicaStatusCode = "IDLE" | "BACKFILL" | "STANDBY" | "PRIMARY" | "FAILED"; +* - BACKFILL: The replica is actively playing the historical recovery log. + - STANDBY: The replica has finished playing the historical recovery log and is +live-tailing it to locally mirror recorded operations as they are +produced. It can take over as primary at any time. + +Shards not having recovery logs immediately transition to STANDBY. + - PRIMARY: The replica is actively serving as primary. + - FAILED: The replica has encountered an unrecoverable error. +*/ +export type ReplicaStatusCode = 'IDLE' | 'BACKFILL' | 'STANDBY' | 'PRIMARY' | 'FAILED'; /** - * Change defines an insertion, update, or deletion to be applied to the set - * of ShardSpecs. Exactly one of |upsert| or |delete| must be set. - */ +* Change defines an insertion, update, or deletion to be applied to the set +of ShardSpecs. Exactly one of |upsert| or |delete| must be set. +*/ export interface ConsumerApplyRequestChange { - /** - * Expected ModRevision of the current ShardSpec. If the shard is being - * created, expect_mod_revision is zero. - * @format int64 - */ - expectModRevision?: string; - /** - * ShardSpec to be updated (if expect_mod_revision > 0) or created - * (if expect_mod_revision == 0). - */ - upsert?: ConsumerShardSpec; - /** Shard to be deleted. expect_mod_revision must not be zero. */ - delete?: string; + /** + * Expected ModRevision of the current ShardSpec. If the shard is being + * created, expect_mod_revision is zero. + * @format int64 + */ + expectModRevision?: string; + /** + * ShardSpec to be updated (if expect_mod_revision > 0) or created + * (if expect_mod_revision == 0). + */ + upsert?: ConsumerShardSpec; + /** Shard to be deleted. expect_mod_revision must not be zero. */ + delete?: string; } export interface ConsumerApplyResponse { - /** Status of the Apply RPC. */ - status?: ConsumerStatus; - /** Header of the response. */ - header?: ProtocolHeader; - /** - * Optional extension of the ApplyResponse. - * @format byte - */ - extension?: string; + /** Status of the Apply RPC. */ + status?: ConsumerStatus; + /** Header of the response. */ + header?: ProtocolHeader; + /** + * Optional extension of the ApplyResponse. + * @format byte + */ + extension?: string; } export interface ConsumerGetHintsResponse { - /** Status of the Hints RPC. */ - status?: ConsumerStatus; - /** Header of the response. */ - header?: ProtocolHeader; - /** Primary hints for the shard. */ - primaryHints?: GetHintsResponseResponseHints; - /** - * List of backup hints for a shard. The most recent recovery log hints will - * be first, any subsequent hints are for historical backup. If there is no - * value for a hint key the value corresponding hints will be nil. - */ - backupHints?: GetHintsResponseResponseHints[]; - /** - * Optional extension of the GetHintsResponse. - * @format byte - */ - extension?: string; + /** Status of the Hints RPC. */ + status?: ConsumerStatus; + /** Header of the response. */ + header?: ProtocolHeader; + /** Primary hints for the shard. */ + primaryHints?: GetHintsResponseResponseHints; + /** + * List of backup hints for a shard. The most recent recovery log hints will + * be first, any subsequent hints are for historical backup. If there is no + * value for a hint key the value corresponding hints will be nil. + */ + backupHints?: GetHintsResponseResponseHints[]; + /** + * Optional extension of the GetHintsResponse. + * @format byte + */ + extension?: string; } export interface ConsumerListRequest { - /** - * Selector optionally refines the set of shards which will be enumerated. - * If zero-valued, all shards are returned. Otherwise, only ShardSpecs - * matching the LabelSelector will be returned. One meta-label "id" is - * additionally supported by the selector, where "id=example-shard-ID" - * will match a ShardSpec with ID "example-shard-ID". - */ - selector?: ProtocolLabelSelector; - /** - * Optional extension of the ListRequest. - * @format byte - */ - extension?: string; + /** + * Selector optionally refines the set of shards which will be enumerated. + * If zero-valued, all shards are returned. Otherwise, only ShardSpecs + * matching the LabelSelector will be returned. One meta-label "id" is + * additionally supported by the selector, where "id=example-shard-ID" + * will match a ShardSpec with ID "example-shard-ID". + */ + selector?: ProtocolLabelSelector; + /** + * Optional extension of the ListRequest. + * @format byte + */ + extension?: string; } export interface ConsumerListResponse { - /** Status of the List RPC. */ - status?: ConsumerStatus; - /** Header of the response. */ - header?: ProtocolHeader; - shards?: ListResponseShard[]; - /** - * Optional extension of the ListResponse. - * @format byte - */ - extension?: string; + /** Status of the List RPC. */ + status?: ConsumerStatus; + /** Header of the response. */ + header?: ProtocolHeader; + shards?: ListResponseShard[]; + /** + * Optional extension of the ListResponse. + * @format byte + */ + extension?: string; } /** - * ReplicaStatus is the status of a ShardSpec assigned to a ConsumerSpec. - * It serves as an allocator AssignmentValue. ReplicaStatus is reduced by taking - * the maximum enum value among statuses. Eg, if a primary is PRIMARY, one - * replica is BACKFILL and the other STANDBY, then the status is PRIMARY. If one - * of the replicas transitioned to FAILED, than the status is FAILED. This - * reduction behavior is used to summarize status across all replicas. - */ +* ReplicaStatus is the status of a ShardSpec assigned to a ConsumerSpec. +It serves as an allocator AssignmentValue. ReplicaStatus is reduced by taking +the maximum enum value among statuses. Eg, if a primary is PRIMARY, one +replica is BACKFILL and the other STANDBY, then the status is PRIMARY. If one +of the replicas transitioned to FAILED, than the status is FAILED. This +reduction behavior is used to summarize status across all replicas. +*/ export interface ConsumerReplicaStatus { - /** - * - BACKFILL: The replica is actively playing the historical recovery log. - * - STANDBY: The replica has finished playing the historical recovery log and is - * live-tailing it to locally mirror recorded operations as they are - * produced. It can take over as primary at any time. - * - * Shards not having recovery logs immediately transition to STANDBY. - * - PRIMARY: The replica is actively serving as primary. - * - FAILED: The replica has encountered an unrecoverable error. - */ - code?: ReplicaStatusCode; - /** Errors encountered during replica processing. Set iff |code| is FAILED. */ - errors?: string[]; + /** + * - BACKFILL: The replica is actively playing the historical recovery log. + * - STANDBY: The replica has finished playing the historical recovery log and is + * live-tailing it to locally mirror recorded operations as they are + * produced. It can take over as primary at any time. + * + * Shards not having recovery logs immediately transition to STANDBY. + * - PRIMARY: The replica is actively serving as primary. + * - FAILED: The replica has encountered an unrecoverable error. + */ + code?: ReplicaStatusCode; + /** Errors encountered during replica processing. Set iff |code| is FAILED. */ + errors?: string[]; } /** - * ShardSpec describes a shard and its configuration, and is the long-lived unit - * of work and scaling for a consumer application. Each shard is allocated to a - * one "primary" at-a-time selected from the current processes of a consumer - * application, and is re-assigned on process fault or exit. - * - * ShardSpecs describe all configuration of the shard and its processing, - * including journals to consume, configuration for processing transactions, its - * recovery log, hot standbys, etc. ShardSpecs may be further extended with - * domain-specific labels & values to further define application behavior. - * ShardSpec is-a allocator.ItemValue. - */ +* ShardSpec describes a shard and its configuration, and is the long-lived unit +of work and scaling for a consumer application. Each shard is allocated to a +one "primary" at-a-time selected from the current processes of a consumer +application, and is re-assigned on process fault or exit. + +ShardSpecs describe all configuration of the shard and its processing, +including journals to consume, configuration for processing transactions, its +recovery log, hot standbys, etc. ShardSpecs may be further extended with +domain-specific labels & values to further define application behavior. +ShardSpec is-a allocator.ItemValue. +*/ export interface ConsumerShardSpec { - /** ID of the shard. */ - id?: string; - /** Sources of the shard, uniquely ordered on Source journal. */ - sources?: ConsumerShardSpecSource[]; - /** - * Prefix of the Journal into which the shard's recovery log will be recorded. - * The complete Journal name is built as "{recovery_log_prefix}/{shard_id}". - * If empty, the shard does not use a recovery log. - */ - recoveryLogPrefix?: string; - /** - * Prefix of Etcd keys into which recovery log FSMHints are written to and - * read from. FSMHints allow readers of the recovery log to efficiently - * determine the minimum fragments of log which must be read to fully recover - * local store state. The complete hint key written by the shard primary is: - * "{hint_prefix}/{shard_id}.primary" - * - * The primary will regularly produce updated hints into this key, and - * players of the log will similarly utilize hints from this key. - * If |recovery_log_prefix| is set, |hint_prefix| must be also. - */ - hintPrefix?: string; - /** - * Backups of verified recovery log FSMHints, retained as a disaster-recovery - * mechanism. On completing playback, a player will write recovered hints to: - * "{hints_prefix}/{shard_id}.backup.0". - * - * It also move hints previously stored under - * "{hints_prefix/{shard_id}.backup.0" to - * "{hints_prefix/{shard_id}.backup.1", and so on, keeping at most - * |hint_backups| distinct sets of FSMHints. - * - * In the case of disaster or data-loss, these copied hints can be an - * important fallback for recovering a consistent albeit older version of the - * shard's store, with each relying on only progressively older portions of - * the recovery log. - * - * When pruning the recovery log, log fragments which are older than (and no - * longer required by) the *oldest* backup are discarded, ensuring that - * all hints remain valid for playback. - * @format int32 - */ - hintBackups?: number; - /** - * Max duration of shard transactions. This duration upper-bounds the amount - * of time during which a transaction may process messages before it must - * flush and commit. It may run for less time if an input message stall occurs - * (eg, no decoded journal message is ready without blocking). A typical value - * would be `1s`: applications which perform extensive aggregation over - * message streams exhibiting locality of "hot" keys may benefit from larger - * values. - */ - maxTxnDuration?: string; - /** - * Min duration of shard transactions. This duration lower-bounds the amount - * of time during which a transaction must process messages before it may - * flush and commit. It may run for more time if additional messages are - * available (eg, decoded journal messages are ready without blocking). Note - * also that transactions are pipelined: a current transaction may process - * messages while a prior transaction's recovery log writes flush to Gazette, - * but it cannot prepare to commit until the prior transaction writes - * complete. In other words even if |min_txn_quantum| is zero, some degree of - * message batching is expected due to the network delay inherent in Gazette - * writes. A typical value of would be `0s`: applications which perform - * extensive aggregation may benefit from larger values. - */ - minTxnDuration?: string; - /** Disable processing of the shard. */ - disable?: boolean; - /** - * Hot standbys is the desired number of consumer processes which should be - * replicating the primary consumer's recovery log. Standbys are allocated in - * a separate availability zone of the current primary, and tail the live log - * to continuously mirror the primary's on-disk DB file structure. Should the - * primary experience failure, one of the hot standbys will be assigned to - * take over as the new shard primary, which is accomplished by simply opening - * its local copy of the recovered store files. - * - * Note that under regular operation, shard hand-off is zero downtime even if - * standbys are zero, as the current primary will not cede ownership until the - * replacement process declares itself ready. However, without standbys a - * process failure will leave the shard without an active primary while its - * replacement starts and completes playback of its recovery log. - * @format int64 - */ - hotStandbys?: number; - /** - * User-defined Labels of this ShardSpec. The label "id" is reserved and may - * not be used with a ShardSpec's labels. - */ - labels?: ProtocolLabelSet; - /** - * Disable waiting for acknowledgements of pending message(s). - * - * If a consumer transaction reads uncommitted messages, it will by default - * remain open (subject to the max duration) awaiting an acknowledgement of - * those messages, in the hope that that acknowledgement will be quickly - * forthcoming and, by remaining open, we can process all messages in this - * transaction. Effectively we're trading a small amount of increased local - * latency for a global reduction in end-to-end latency. - * - * This works well for acyclic message flows, but can introduce unnecessary - * stalls if there are message cycles between shards. In the simplest case, - * a transaction could block awaiting an ACK of a message that it itself - * produced -- an ACK which can't arrive until the transaction closes. - */ - disableWaitForAck?: boolean; - /** - * Size of the ring buffer used to sequence read-uncommitted messages - * into consumed, read-committed ones. The ring buffer is a performance - * optimization only: applications will replay portions of journals as - * needed when messages aren't available in the buffer. - * It can remain small if source journal transactions are small, - * but larger transactions will achieve better performance with a - * larger ring. - * If zero, a reasonable default (currently 8192) is used. - * @format int64 - */ - ringBufferSize?: number; - /** - * Size of the channel used to bridge message read and decode with - * sequencing and consumption. Larger values may reduce data stalls, - * particularly for larger transactions and/or bursty custom - * MessageProducer implementations. - * If zero, a reasonable default (currently 8192) is used. - * @format int64 - */ - readChannelSize?: number; + /** ID of the shard. */ + id?: string; + /** Sources of the shard, uniquely ordered on Source journal. */ + sources?: ConsumerShardSpecSource[]; + /** + * Prefix of the Journal into which the shard's recovery log will be recorded. + * The complete Journal name is built as "{recovery_log_prefix}/{shard_id}". + * If empty, the shard does not use a recovery log. + */ + recoveryLogPrefix?: string; + /** + * Prefix of Etcd keys into which recovery log FSMHints are written to and + * read from. FSMHints allow readers of the recovery log to efficiently + * determine the minimum fragments of log which must be read to fully recover + * local store state. The complete hint key written by the shard primary is: + * "{hint_prefix}/{shard_id}.primary" + * + * The primary will regularly produce updated hints into this key, and + * players of the log will similarly utilize hints from this key. + * If |recovery_log_prefix| is set, |hint_prefix| must be also. + */ + hintPrefix?: string; + /** + * Backups of verified recovery log FSMHints, retained as a disaster-recovery + * mechanism. On completing playback, a player will write recovered hints to: + * "{hints_prefix}/{shard_id}.backup.0". + * + * It also move hints previously stored under + * "{hints_prefix/{shard_id}.backup.0" to + * "{hints_prefix/{shard_id}.backup.1", and so on, keeping at most + * |hint_backups| distinct sets of FSMHints. + * In the case of disaster or data-loss, these copied hints can be an + * important fallback for recovering a consistent albeit older version of the + * shard's store, with each relying on only progressively older portions of + * the recovery log. + * When pruning the recovery log, log fragments which are older than (and no + * longer required by) the *oldest* backup are discarded, ensuring that + * all hints remain valid for playback. + * @format int32 + */ + hintBackups?: number; + /** + * Max duration of shard transactions. This duration upper-bounds the amount + * of time during which a transaction may process messages before it must + * flush and commit. It may run for less time if an input message stall occurs + * (eg, no decoded journal message is ready without blocking). A typical value + * would be `1s`: applications which perform extensive aggregation over + * message streams exhibiting locality of "hot" keys may benefit from larger + * values. + */ + maxTxnDuration?: string; + /** + * Min duration of shard transactions. This duration lower-bounds the amount + * of time during which a transaction must process messages before it may + * flush and commit. It may run for more time if additional messages are + * available (eg, decoded journal messages are ready without blocking). Note + * also that transactions are pipelined: a current transaction may process + * messages while a prior transaction's recovery log writes flush to Gazette, + * but it cannot prepare to commit until the prior transaction writes + * complete. In other words even if |min_txn_quantum| is zero, some degree of + * message batching is expected due to the network delay inherent in Gazette + * writes. A typical value of would be `0s`: applications which perform + * extensive aggregation may benefit from larger values. + */ + minTxnDuration?: string; + /** Disable processing of the shard. */ + disable?: boolean; + /** + * Hot standbys is the desired number of consumer processes which should be + * replicating the primary consumer's recovery log. Standbys are allocated in + * a separate availability zone of the current primary, and tail the live log + * to continuously mirror the primary's on-disk DB file structure. Should the + * primary experience failure, one of the hot standbys will be assigned to + * take over as the new shard primary, which is accomplished by simply opening + * its local copy of the recovered store files. + * + * Note that under regular operation, shard hand-off is zero downtime even if + * standbys are zero, as the current primary will not cede ownership until the + * replacement process declares itself ready. However, without standbys a + * process failure will leave the shard without an active primary while its + * replacement starts and completes playback of its recovery log. + * @format int64 + */ + hotStandbys?: number; + /** + * User-defined Labels of this ShardSpec. The label "id" is reserved and may + * not be used with a ShardSpec's labels. + */ + labels?: ProtocolLabelSet; + /** + * Disable waiting for acknowledgements of pending message(s). + * + * If a consumer transaction reads uncommitted messages, it will by default + * remain open (subject to the max duration) awaiting an acknowledgement of + * those messages, in the hope that that acknowledgement will be quickly + * forthcoming and, by remaining open, we can process all messages in this + * transaction. Effectively we're trading a small amount of increased local + * latency for a global reduction in end-to-end latency. + * This works well for acyclic message flows, but can introduce unnecessary + * stalls if there are message cycles between shards. In the simplest case, + * a transaction could block awaiting an ACK of a message that it itself + * produced -- an ACK which can't arrive until the transaction closes. + */ + disableWaitForAck?: boolean; + /** + * Size of the ring buffer used to sequence read-uncommitted messages + * into consumed, read-committed ones. The ring buffer is a performance + * optimization only: applications will replay portions of journals as + * needed when messages aren't available in the buffer. + * It can remain small if source journal transactions are small, + * but larger transactions will achieve better performance with a + * larger ring. + * If zero, a reasonable default (currently 8192) is used. + * @format int64 + */ + ringBufferSize?: number; + /** + * Size of the channel used to bridge message read and decode with + * sequencing and consumption. Larger values may reduce data stalls, + * particularly for larger transactions and/or bursty custom + * MessageProducer implementations. + * If zero, a reasonable default (currently 8192) is used. + * @format int64 + */ + readChannelSize?: number; } /** - * Sources define the set of journals which this shard consumes. At least one - * Source must be specified, and in many use cases only one will be needed. - * For use cases which can benefit, multiple sources may be specified to - * represent a "join" over messages of distinct journals. - * - * Note the effective mapping of messages to each of the joined journals - * should align (eg, joining a journal of customer updates with one of orders, - * where both are mapped on customer ID). This typically means the - * partitioning of the two event "topics" must be the same. - * - * Another powerful pattern is to shard on partitions of a high-volume event - * stream, and also have each shard join against all events of a low-volume - * stream. For example, a shard might ingest and index "viewed product" - * events, read a comparably low-volume "purchase" event stream, and on each - * purchase publish the bundle of its corresponding prior product views. - */ +* Sources define the set of journals which this shard consumes. At least one +Source must be specified, and in many use cases only one will be needed. +For use cases which can benefit, multiple sources may be specified to +represent a "join" over messages of distinct journals. + +Note the effective mapping of messages to each of the joined journals +should align (eg, joining a journal of customer updates with one of orders, +where both are mapped on customer ID). This typically means the +partitioning of the two event "topics" must be the same. + +Another powerful pattern is to shard on partitions of a high-volume event +stream, and also have each shard join against all events of a low-volume +stream. For example, a shard might ingest and index "viewed product" +events, read a comparably low-volume "purchase" event stream, and on each +purchase publish the bundle of its corresponding prior product views. +*/ export interface ConsumerShardSpecSource { - /** Journal which this shard is consuming. */ - journal?: string; - /** - * Minimum journal byte offset the shard should begin reading from. - * Typically this should be zero, as read offsets are check-pointed and - * restored from the shard's Store as it processes. |min_offset| can be - * useful for shard initialization, directing it to skip over historical - * portions of the journal not needed for the application's use case. - * @format int64 - */ - minOffset?: string; + /** Journal which this shard is consuming. */ + journal?: string; + /** + * Minimum journal byte offset the shard should begin reading from. + * Typically this should be zero, as read offsets are check-pointed and + * restored from the shard's Store as it processes. |min_offset| can be + * useful for shard initialization, directing it to skip over historical + * portions of the journal not needed for the application's use case. + * @format int64 + */ + minOffset?: string; } export interface ConsumerStatRequest { - /** Header may be attached by a proxying consumer peer. */ - header?: ProtocolHeader; - /** Shard to Stat. */ - shard?: string; - /** - * Journals and offsets which must be reflected in a completed consumer - * transaction before Stat returns, blocking if required. Offsets of journals - * not read by this shard are ignored. - */ - readThrough?: Record; - /** - * Optional extension of the StatRequest. - * @format byte - */ - extension?: string; + /** Header may be attached by a proxying consumer peer. */ + header?: ProtocolHeader; + /** Shard to Stat. */ + shard?: string; + /** + * Journals and offsets which must be reflected in a completed consumer + * transaction before Stat returns, blocking if required. Offsets of journals + * not read by this shard are ignored. + */ + readThrough?: Record; + /** + * Optional extension of the StatRequest. + * @format byte + */ + extension?: string; } export interface ConsumerStatResponse { - /** Status of the Stat RPC. */ - status?: ConsumerStatus; - /** Header of the response. */ - header?: ProtocolHeader; - /** - * Journals and offsets read through by the most recent completed consumer - * transaction. - */ - readThrough?: Record; - /** - * Journals and offsets this shard has published through, including - * acknowledgements, as-of the most recent completed consumer transaction. - * - * Formally, if an acknowledged message A results in this shard publishing - * messages B, and A falls within |read_through|, then all messages B & their - * acknowledgements fall within |publish_at|. - * - * The composition of |read_through| and |publish_at| allow CQRS applications - * to provide read-your-writes consistency, even if written events pass - * through multiple intermediate consumers and arbitrary transformations - * before arriving at the materialized view which is ultimately queried. - */ - publishAt?: Record; - /** - * Optional extension of the StatResponse. - * @format byte - */ - extension?: string; + /** Status of the Stat RPC. */ + status?: ConsumerStatus; + /** Header of the response. */ + header?: ProtocolHeader; + /** + * Journals and offsets read through by the most recent completed consumer + * transaction. + */ + readThrough?: Record; + /** + * Journals and offsets this shard has published through, including + * acknowledgements, as-of the most recent completed consumer transaction. + * + * Formally, if an acknowledged message A results in this shard publishing + * messages B, and A falls within |read_through|, then all messages B & their + * acknowledgements fall within |publish_at|. + * The composition of |read_through| and |publish_at| allow CQRS applications + * to provide read-your-writes consistency, even if written events pass + * through multiple intermediate consumers and arbitrary transformations + * before arriving at the materialized view which is ultimately queried. + */ + publishAt?: Record; + /** + * Optional extension of the StatResponse. + * @format byte + */ + extension?: string; } /** - * Status is a response status code, used across Gazette Consumer RPC APIs. - * - * - SHARD_NOT_FOUND: The named shard does not exist. - * - NO_SHARD_PRIMARY: There is no current primary consumer process for the shard. This is a - * temporary condition which should quickly resolve, assuming sufficient - * consumer capacity. - * - NOT_SHARD_PRIMARY: The present consumer process is not the assigned primary for the shard, - * and was not instructed to proxy the request. - * - ETCD_TRANSACTION_FAILED: The Etcd transaction failed. Returned by Update RPC when an - * expect_mod_revision of the UpdateRequest differs from the current - * ModRevision of the ShardSpec within the store. - * - SHARD_STOPPED: The current primary shard has stopped, either due to reassignment or - * processing failure, and will not make further progress toward the - * requested operation. - * For example, a Stat RPC will return SHARD_STOPPED if the StatRequest - * cannot be satisfied. - * @default "OK" - */ +* Status is a response status code, used across Gazette Consumer RPC APIs. + + - SHARD_NOT_FOUND: The named shard does not exist. + - NO_SHARD_PRIMARY: There is no current primary consumer process for the shard. This is a +temporary condition which should quickly resolve, assuming sufficient +consumer capacity. + - NOT_SHARD_PRIMARY: The present consumer process is not the assigned primary for the shard, +and was not instructed to proxy the request. + - ETCD_TRANSACTION_FAILED: The Etcd transaction failed. Returned by Update RPC when an +expect_mod_revision of the UpdateRequest differs from the current +ModRevision of the ShardSpec within the store. + - SHARD_STOPPED: The current primary shard has stopped, either due to reassignment or +processing failure, and will not make further progress toward the +requested operation. +For example, a Stat RPC will return SHARD_STOPPED if the StatRequest +cannot be satisfied. +*/ export type ConsumerStatus = - | "OK" - | "SHARD_NOT_FOUND" - | "NO_SHARD_PRIMARY" - | "NOT_SHARD_PRIMARY" - | "ETCD_TRANSACTION_FAILED" - | "SHARD_STOPPED"; + | 'OK' + | 'SHARD_NOT_FOUND' + | 'NO_SHARD_PRIMARY' + | 'NOT_SHARD_PRIMARY' + | 'ETCD_TRANSACTION_FAILED' + | 'SHARD_STOPPED'; export interface ConsumerUnassignResponse { - /** Status of the Unassign RPC. */ - status?: ConsumerStatus; - /** Shards which had assignments removed. */ - shards?: string[]; + /** Status of the Unassign RPC. */ + status?: ConsumerStatus; + /** Shards which had assignments removed. */ + shards?: string[]; } export interface ProtobufAny { - typeUrl?: string; - /** @format byte */ - value?: string; + typeUrl?: string; + /** @format byte */ + value?: string; } /** - * Header captures metadata such as the process responsible for processing - * an RPC, and its effective Etcd state. - */ +* Header captures metadata such as the process responsible for processing +an RPC, and its effective Etcd state. +*/ export interface ProtocolHeader { - /** - * ID of the process responsible for request processing. May be empty iff - * Header is being used within a proxied request, and that request may be - * dispatched to any member of the Route. - */ - processId?: ProcessSpecID; - /** - * Route of processes specifically responsible for this RPC, or an empty Route - * if any process is capable of serving the RPC. - */ - route?: ProtocolRoute; - /** - * Etcd represents the effective Etcd MVCC state under which a Gazette broker - * is operating in its processing of requests and responses. Its inclusion - * allows brokers to reason about relative "happened before" Revision ordering - * of apparent routing conflicts in proxied or replicated requests, as well - * as enabling sanity checks over equality of Etcd ClusterId (and precluding, - * for example, split-brain scenarios where different brokers are backed by - * different Etcd clusters). Etcd is kept in sync with - * etcdserverpb.ResponseHeader. - */ - etcd?: HeaderEtcd; + /** + * ID of the process responsible for request processing. May be empty iff + * Header is being used within a proxied request, and that request may be + * dispatched to any member of the Route. + */ + processId?: ProcessSpecID; + /** + * Route of processes specifically responsible for this RPC, or an empty Route + * if any process is capable of serving the RPC. + */ + route?: ProtocolRoute; + /** + * Etcd represents the effective Etcd MVCC state under which a Gazette broker + * is operating in its processing of requests and responses. Its inclusion + * allows brokers to reason about relative "happened before" Revision ordering + * of apparent routing conflicts in proxied or replicated requests, as well + * as enabling sanity checks over equality of Etcd ClusterId (and precluding, + * for example, split-brain scenarios where different brokers are backed by + * different Etcd clusters). Etcd is kept in sync with + * etcdserverpb.ResponseHeader. + */ + etcd?: HeaderEtcd; } /** - * Label defines a key & value pair which can be attached to entities like - * JournalSpecs and BrokerSpecs. Labels may be used to provide identifying - * attributes which do not directly imply semantics to the core system, but - * are meaningful to users or for higher-level Gazette tools. - */ +* Label defines a key & value pair which can be attached to entities like +JournalSpecs and BrokerSpecs. Labels may be used to provide identifying +attributes which do not directly imply semantics to the core system, but +are meaningful to users or for higher-level Gazette tools. +*/ export interface ProtocolLabel { - name?: string; - value?: string; + name?: string; + value?: string; } -/** LabelSelector defines a filter over LabelSets. */ +/** + * LabelSelector defines a filter over LabelSets. + */ export interface ProtocolLabelSelector { - /** - * Include is Labels which must be matched for a LabelSet to be selected. If - * empty, all Labels are included. An include Label with empty ("") value is - * matched by a Label of the same name having any value. - */ - include?: ProtocolLabelSet; - /** - * Exclude is Labels which cannot be matched for a LabelSet to be selected. If - * empty, no Labels are excluded. An exclude Label with empty ("") value - * excludes a Label of the same name having any value. - */ - exclude?: ProtocolLabelSet; + /** + * Include is Labels which must be matched for a LabelSet to be selected. If + * empty, all Labels are included. An include Label with empty ("") value is + * matched by a Label of the same name having any value. + */ + include?: ProtocolLabelSet; + /** + * Exclude is Labels which cannot be matched for a LabelSet to be selected. If + * empty, no Labels are excluded. An exclude Label with empty ("") value + * excludes a Label of the same name having any value. + */ + exclude?: ProtocolLabelSet; } -/** LabelSet is a collection of labels and their values. */ +/** + * LabelSet is a collection of labels and their values. + */ export interface ProtocolLabelSet { - /** Labels of the set. Instances must be unique and sorted over (Name, Value). */ - labels?: ProtocolLabel[]; + /** Labels of the set. Instances must be unique and sorted over (Name, Value). */ + labels?: ProtocolLabel[]; } -/** Route captures the current topology of an item and the processes serving it. */ +/** + * Route captures the current topology of an item and the processes serving it. + */ export interface ProtocolRoute { - /** Members of the Route, ordered on ascending ProcessSpec.ID (zone, suffix). */ - members?: ProcessSpecID[]; - /** - * Index of the ProcessSpec serving as primary within |members|, - * or -1 of no member is currently primary. - * @format int32 - */ - primary?: number; - /** - * Endpoints of each Route member. If not empty, |endpoints| has the same - * length and order as |members|, and captures the endpoint of each one. - */ - endpoints?: string[]; + /** Members of the Route, ordered on ascending ProcessSpec.ID (zone, suffix). */ + members?: ProcessSpecID[]; + /** + * Index of the ProcessSpec serving as primary within |members|, + * or -1 of no member is currently primary. + * @format int32 + */ + primary?: number; + /** + * Endpoints of each Route member. If not empty, |endpoints| has the same + * length and order as |members|, and captures the endpoint of each one. + */ + endpoints?: string[]; } /** - * FSMHints represents a manifest of Fnodes which were still live (eg, having - * remaining links) at the time the FSMHints were produced, as well as any - * Properties. It allows a Player of the log to identify minimal Segments which - * must be read to recover all Fnodes, and also contains sufficient metadata for - * a Player to resolve all possible conflicts it could encounter while reading - * the log, to arrive at a consistent view of file state which exactly matches - * that of the Recorder producing the FSMHints. - * Next tag: 4. - */ +* FSMHints represents a manifest of Fnodes which were still live (eg, having +remaining links) at the time the FSMHints were produced, as well as any +Properties. It allows a Player of the log to identify minimal Segments which +must be read to recover all Fnodes, and also contains sufficient metadata for +a Player to resolve all possible conflicts it could encounter while reading +the log, to arrive at a consistent view of file state which exactly matches +that of the Recorder producing the FSMHints. +Next tag: 4. +*/ export interface RecoverylogFSMHints { - /** - * Log is the implied recovery log of any contained |live_nodes| Segments - * which omit a |log| value. This implied behavior is both for backward- - * compatibility (Segments didn't always have a |log| field) and also for - * compacting the representation in the common case of Segments mostly or - * entirely addressing a single log. - */ - log?: string; - /** Live Fnodes and their Segments as-of the generation of these FSMHints. */ - liveNodes?: RecoverylogFnodeSegments[]; - /** Property files and contents as-of the generation of these FSMHints. */ - properties?: RecoverylogProperty[]; + /** + * Log is the implied recovery log of any contained |live_nodes| Segments + * which omit a |log| value. This implied behavior is both for backward- + * compatibility (Segments didn't always have a |log| field) and also for + * compacting the representation in the common case of Segments mostly or + * entirely addressing a single log. + */ + log?: string; + /** Live Fnodes and their Segments as-of the generation of these FSMHints. */ + liveNodes?: RecoverylogFnodeSegments[]; + /** Property files and contents as-of the generation of these FSMHints. */ + properties?: RecoverylogProperty[]; } -/** FnodeSegments captures log Segments containing all RecordedOps of the Fnode. */ +/** + * FnodeSegments captures log Segments containing all RecordedOps of the Fnode. + */ export interface RecoverylogFnodeSegments { - /** - * Fnode being hinted. - * @format int64 - */ - fnode?: string; - /** - * Segments of the Fnode in the log. Currently, FSM tracks only a single - * Segment per Fnode per Author & Log. A specific implication of this is that Fnodes - * modified over long periods of time will result in Segments spanning large - * chunks of the log. For best performance, Fnodes should be opened & written - * once, and then never be modified again (this is RocksDB's behavior). - * If supporting this case is desired, FSM will have to be a bit smarter about - * not extending Segments which gap over significant portions of the log - * (eg, there's a trade-off to make over size of the hinted manifest, vs - * savings incurred on playback by being able to skip portions of the log). - */ - segments?: RecoverylogSegment[]; + /** + * Fnode being hinted. + * @format int64 + */ + fnode?: string; + /** + * Segments of the Fnode in the log. Currently, FSM tracks only a single + * Segment per Fnode per Author & Log. A specific implication of this is that Fnodes + * modified over long periods of time will result in Segments spanning large + * chunks of the log. For best performance, Fnodes should be opened & written + * once, and then never be modified again (this is RocksDB's behavior). + * If supporting this case is desired, FSM will have to be a bit smarter about + * not extending Segments which gap over significant portions of the log + * (eg, there's a trade-off to make over size of the hinted manifest, vs + * savings incurred on playback by being able to skip portions of the log). + */ + segments?: RecoverylogSegment[]; } /** - * Property is a small file which rarely changes, and is thus managed - * outside of regular Fnode tracking. See FSM.Properties. - */ +* Property is a small file which rarely changes, and is thus managed +outside of regular Fnode tracking. See FSM.Properties. +*/ export interface RecoverylogProperty { - /** Filesystem path of this property, relative to the common base directory. */ - path?: string; - /** Complete file content of this property. */ - content?: string; + /** Filesystem path of this property, relative to the common base directory. */ + path?: string; + /** Complete file content of this property. */ + content?: string; } /** - * Segment is a contiguous chunk of recovery log written by a single Author. - * Recorders track Segments they have written, for use in providing hints to - * future readers of the log. A key point to understand is that Gazette append - * semantics mean that Recorders *cannot know* exactly what offsets their writes - * are applied to in the log, nor guarantee that their operations are not being - * interleaved with those of other writers. Log Players are aware of these - * limitations, and use Segments to resolve conflicts of possible interpretation - * of the log. Segments produced by a Player are exact, since Players observe all - * recorded operations at their exact offsets. - * Next tag: 8. - */ +* Segment is a contiguous chunk of recovery log written by a single Author. +Recorders track Segments they have written, for use in providing hints to +future readers of the log. A key point to understand is that Gazette append +semantics mean that Recorders *cannot know* exactly what offsets their writes +are applied to in the log, nor guarantee that their operations are not being +interleaved with those of other writers. Log Players are aware of these +limitations, and use Segments to resolve conflicts of possible interpretation +of the log. Segments produced by a Player are exact, since Players observe all +recorded operations at their exact offsets. +Next tag: 8. +*/ export interface RecoverylogSegment { - /** - * Author which wrote RecordedOps of this Segment. - * @format int64 - */ - author?: number; - /** - * First (lowest) sequence number of RecordedOps within this Segment. - * @format int64 - */ - firstSeqNo?: string; - /** - * First byte offset of the Segment, where |first_seq_no| is recorded. - * If this Segment was produced by a Recorder, this is guaranteed only to be a - * lower-bound (eg, a Player reading at this offset may encounter irrelevant - * operations prior to the RecordedOp indicated by the tuple - * (|author|, |first_seq_no|, |first_checksum|). If a Player produced the Segment, - * first_offset is exact. - * @format int64 - */ - firstOffset?: string; - /** - * Checksum of the RecordedOp having |first_seq_no|. - * @format int64 - */ - firstChecksum?: number; - /** - * Last (highest, inclusive) sequence number of RecordedOps within this Segment. - * @format int64 - */ - lastSeqNo?: string; - /** - * Last offset (exclusive) of the Segment. Zero means the offset is not known - * (eg, because the Segment was produced by a Recorder). - * @format int64 - */ - lastOffset?: string; - /** Log is the Journal holding this Segment's data, and to which offsets are relative. */ - log?: string; + /** + * Author which wrote RecordedOps of this Segment. + * @format int64 + */ + author?: number; + /** + * First (lowest) sequence number of RecordedOps within this Segment. + * @format int64 + */ + firstSeqNo?: string; + /** + * First byte offset of the Segment, where |first_seq_no| is recorded. + * If this Segment was produced by a Recorder, this is guaranteed only to be a + * lower-bound (eg, a Player reading at this offset may encounter irrelevant + * operations prior to the RecordedOp indicated by the tuple + * (|author|, |first_seq_no|, |first_checksum|). If a Player produced the Segment, + * first_offset is exact. + * @format int64 + */ + firstOffset?: string; + /** + * Checksum of the RecordedOp having |first_seq_no|. + * @format int64 + */ + firstChecksum?: number; + /** + * Last (highest, inclusive) sequence number of RecordedOps within this Segment. + * @format int64 + */ + lastSeqNo?: string; + /** + * Last offset (exclusive) of the Segment. Zero means the offset is not known + * (eg, because the Segment was produced by a Recorder). + * @format int64 + */ + lastOffset?: string; + /** Log is the Journal holding this Segment's data, and to which offsets are relative. */ + log?: string; } export interface RuntimeError { - error?: string; - /** @format int32 */ - code?: number; - message?: string; - details?: ProtobufAny[]; + error?: string; + /** @format int32 */ + code?: number; + message?: string; + details?: ProtobufAny[]; } diff --git a/gen/broker/protocol/protocol.swagger.json b/gen/broker/protocol/protocol.swagger.json index 506c3b7..f61d1ba 100644 --- a/gen/broker/protocol/protocol.swagger.json +++ b/gen/broker/protocol/protocol.swagger.json @@ -504,6 +504,11 @@ "type": "string", "format": "int64", "description": "Offset to read through. If zero, then the read end offset is unconstrained." + }, + "beginModTime": { + "type": "string", + "format": "int64", + "description": "BeginModTime is an optional inclusive lower bound on the modification\ntimestamps of fragments read from the backing store, represented as\nseconds since the epoch. The request Offset will be advanced as-needed\nto skip persisted Fragments having a modication time before the bound." } }, "description": "ReadRequest is the unary request message of the broker Read RPC." diff --git a/go.mod b/go.mod index 35178f0..5bc19b2 100644 --- a/go.mod +++ b/go.mod @@ -5,7 +5,7 @@ go 1.19 require ( github.com/estuary/flow v0.1.9-0.20230303181027-f65a9d7f1a89 github.com/gogo/gateway v1.1.0 - github.com/golang-jwt/jwt/v4 v4.4.1 + github.com/golang-jwt/jwt/v4 v4.5.0 github.com/golang/protobuf v1.5.2 github.com/grpc-ecosystem/go-grpc-prometheus v1.2.0 github.com/grpc-ecosystem/grpc-gateway v1.16.0 @@ -15,7 +15,7 @@ require ( github.com/sirupsen/logrus v1.9.0 github.com/urfave/negroni v1.0.0 github.com/valyala/fasthttp v1.48.0 - go.gazette.dev/core v0.89.1-0.20230106205712-a2d374cdd6be + go.gazette.dev/core v0.89.1-0.20231012132739-dfed675b7fd1 golang.org/x/net v0.8.0 golang.org/x/sync v0.0.0-20220601150217-0de741cfad7f google.golang.org/grpc v1.53.0 diff --git a/go.sum b/go.sum index 488e89f..878b585 100644 --- a/go.sum +++ b/go.sum @@ -13,17 +13,6 @@ cloud.google.com/go v0.56.0/go.mod h1:jr7tqZxxKOVYizybht9+26Z/gUq7tiRzu+ACVAMbKV cloud.google.com/go v0.57.0/go.mod h1:oXiQ6Rzq3RAkkY7N6t3TcE6jE+CIBBbA36lwQ1JyzZs= cloud.google.com/go v0.62.0/go.mod h1:jmCYTdRCQuc1PHIIJ/maLInMho30T/Y0M4hTdTShOYc= cloud.google.com/go v0.65.0/go.mod h1:O5N8zS7uWy9vkA9vayVHs65eM1ubvY4h553ofrNHObY= -cloud.google.com/go v0.72.0/go.mod h1:M+5Vjvlc2wnp6tjzE102Dw08nGShTscUx2nZMufOKPI= -cloud.google.com/go v0.74.0/go.mod h1:VV1xSbzvo+9QJOxLDaJfTjx5e+MePCpCWwvftOeQmWk= -cloud.google.com/go v0.78.0/go.mod h1:QjdrLG0uq+YwhjoVOLsS1t7TW8fs36kLs4XO5R5ECHg= -cloud.google.com/go v0.79.0/go.mod h1:3bzgcEeQlzbuEAYu4mrWhKqWjmpprinYgKJLgKHnbb8= -cloud.google.com/go v0.81.0/go.mod h1:mk/AM35KwGk/Nm2YSeZbxXdrNK3KZOYHmLkOqC2V6E0= -cloud.google.com/go v0.83.0/go.mod h1:Z7MJUsANfY0pYPdw0lbnivPx4/vhy/e2FEkSkF7vAVY= -cloud.google.com/go v0.84.0/go.mod h1:RazrYuxIK6Kb7YrzzhPoLmCVzl7Sup4NrbKPg8KHSUM= -cloud.google.com/go v0.87.0/go.mod h1:TpDYlFy7vuLzZMMZ+B6iRiELaY7z/gJPaqbMx6mlWcY= -cloud.google.com/go v0.90.0/go.mod h1:kRX0mNRHe0e2rC6oNakvwQqzyDmg57xJ+SZU1eT2aDQ= -cloud.google.com/go v0.93.3/go.mod h1:8utlLll2EF5XMAV15woO4lSbWQlk8rer9aLOfLh7+YI= -cloud.google.com/go v0.94.0/go.mod h1:qAlAugsXlC+JWO+Bke5vCtc9ONxjQT3drlTTnAplMW4= cloud.google.com/go v0.105.0 h1:DNtEKRBAAzeS4KyIory52wWHuClNaXJ5x1F7xa4q+5Y= cloud.google.com/go/bigquery v1.0.1/go.mod h1:i/xbL2UlR5RvWAURpBYZTtm/cXjCha9lbfbpx4poX+o= cloud.google.com/go/bigquery v1.3.0/go.mod h1:PjpwJnslEMmckchkHFfq+HTD2DmtT67aNFKH1/VBDHE= @@ -45,25 +34,19 @@ cloud.google.com/go/storage v1.5.0/go.mod h1:tpKbwo567HUNpVclU5sGELwQWBDZ8gh0Zeo cloud.google.com/go/storage v1.6.0/go.mod h1:N7U0C8pVQ/+NIKOBQyamJIeKQKkZ+mxpohlUTyfDhBk= cloud.google.com/go/storage v1.8.0/go.mod h1:Wv1Oy7z6Yz3DshWRJFhqM/UCfaWIRTdp0RXyy7KQOVs= cloud.google.com/go/storage v1.10.0/go.mod h1:FLPqc6j+Ki4BU591ie1oL6qBQGu2Bl/tZ9ullr3+Kg0= -cloud.google.com/go/storage v1.16.1/go.mod h1:LaNorbty3ehnU3rEjXSNV/NRgQA0O8Y+uh6bPe5UOk4= cloud.google.com/go/storage v1.18.2 h1:5NQw6tOn3eMm0oE8vTkfjau18kjL79FlMjy/CHTpmoY= dmitri.shuralyov.com/gpu/mtl v0.0.0-20190408044501-666a987793e9/go.mod h1:H6x//7gZCb22OMCxBHrMx7a5I7Hp++hsVxbQ4BYO7hU= github.com/Azure/azure-pipeline-go v0.2.3 h1:7U9HBg1JFK3jHl5qmo4CTZKFTVgMwdFHMVtCdfBE21U= -github.com/Azure/azure-pipeline-go v0.2.3/go.mod h1:x841ezTBIMG6O3lAcl8ATHnsOPVl2bqk7S3ta6S6u4k= -github.com/Azure/azure-storage-blob-go v0.14.0 h1:1BCg74AmVdYwO3dlKwtFU1V0wU2PZdREkXvAmZJRUlM= -github.com/Azure/azure-storage-blob-go v0.14.0/go.mod h1:SMqIBi+SuiQH32bvyjngEewEeXoPfKMgWlBDaYf6fck= -github.com/Azure/go-autorest v11.1.2+incompatible/go.mod h1:r+4oMnoxhatjLLJ6zxSWATqVooLgysK6ZNox3g/xq24= -github.com/Azure/go-autorest v14.2.0+incompatible/go.mod h1:r+4oMnoxhatjLLJ6zxSWATqVooLgysK6ZNox3g/xq24= -github.com/Azure/go-autorest/autorest/adal v0.9.13/go.mod h1:W/MM4U6nLxnIskrw4UwWzlHfGjwUS50aOsc/I3yuU8M= -github.com/Azure/go-autorest/autorest/date v0.3.0/go.mod h1:BI0uouVdmngYNUzGWeSYnokU+TrmwEsOqdt8Y6sso74= -github.com/Azure/go-autorest/autorest/mocks v0.4.1/go.mod h1:LTp+uSrOhSkaKrUy935gNZuuIPPVsHlr9DSOxSayd+k= -github.com/Azure/go-autorest/logger v0.2.1/go.mod h1:T9E3cAhj2VqvPOtCYAvby9aBXkZmbF5NWuPV8+WeEW8= -github.com/Azure/go-autorest/tracing v0.6.0/go.mod h1:+vhtPC754Xsa23ID7GlGsrdKBpUA79WCAKPPZVC2DeU= +github.com/Azure/azure-sdk-for-go/sdk/azcore v1.6.1 h1:SEy2xmstIphdPwNBUi7uhvjyjhVKISfwjfOJmuy7kg4= +github.com/Azure/azure-sdk-for-go/sdk/azidentity v1.3.0 h1:vcYCAze6p19qBW7MhZybIsqD8sMV8js0NyQM8JDnVtg= +github.com/Azure/azure-sdk-for-go/sdk/internal v1.3.0 h1:sXr+ck84g/ZlZUOZiNELInmMgOsuGwdjjVkEIde0OtY= +github.com/Azure/azure-sdk-for-go/sdk/storage/azblob v1.0.0 h1:u/LLAOFgsMv7HmNL4Qufg58y+qElGOt5qv0z1mURkRY= +github.com/Azure/azure-storage-blob-go v0.15.0 h1:rXtgp8tN1p29GvpGgfJetavIG0V7OgcSXPpwp3tx6qk= +github.com/AzureAD/microsoft-authentication-library-for-go v1.0.0 h1:OBhqkivkhkMqLPymWEppkm7vgPQY2XsHoEkaMQ0AdZY= github.com/BurntSushi/toml v0.3.1/go.mod h1:xHWCNGjB5oqiDr8zfno3MHue2Ht5sIBksp03qcyfWMU= github.com/BurntSushi/xgb v0.0.0-20160522181843-27f122750802/go.mod h1:IVnqGOEym/WlBOVXweHU+Q+/VP0lqqI8lqeDx9IjBqo= github.com/DataDog/zstd v1.4.8 h1:Rpmta4xZ/MgZnriKNd24iZMhGpP5dvUcs/uqfBapKZY= github.com/DataDog/zstd v1.4.8/go.mod h1:g4AWEaM3yOg3HYfnJ3YIawPnVdXJh9QME85blwSAmyw= -github.com/OneOfOne/xxhash v1.2.2/go.mod h1:HSdplMjZKSmBqAxg5vPj2TmRDmfkzw+cTzAElWljhcU= github.com/alecthomas/template v0.0.0-20160405071501-a0175ee3bccc/go.mod h1:LOuyumcjzFXgccqObfd/Ljyb9UuFJ6TxHnclSeseNhc= github.com/alecthomas/template v0.0.0-20190718012654-fb15b899a751/go.mod h1:LOuyumcjzFXgccqObfd/Ljyb9UuFJ6TxHnclSeseNhc= github.com/alecthomas/units v0.0.0-20151022065526-2efee857e7cf/go.mod h1:ybxpYRFXyAe+OPACYpWeL0wqObRcbAqCMya13uyzqw0= @@ -73,15 +56,12 @@ github.com/andybalholm/brotli v1.0.5 h1:8uQZIdzKmjc/iuPu7O2ioW48L81FgatrcpfFmiq/ github.com/andybalholm/brotli v1.0.5/go.mod h1:fO7iG3H7G2nSZ7m0zPUDn85XEX2GTukHGRSepvi9Eig= github.com/antihax/optional v1.0.0/go.mod h1:uupD/76wgC+ih3iEmQUL+0Ugr19nfwCT1kdvxnR2qWY= github.com/aws/aws-sdk-go v1.40.35 h1:ofWh1LlWaSbOpAsl8EHlg96PZXqgCGKKi8YgrdU2Z+I= -github.com/aws/aws-sdk-go v1.40.35/go.mod h1:585smgzpB/KqRA+K3y/NL/oYRqQvpNJYvLm+LY1U59Q= -github.com/benbjohnson/clock v1.1.0/go.mod h1:J11/hYXuz8f4ySSvYwY0FKfm+ezbsZBKZxNJlLklBHA= github.com/beorn7/perks v0.0.0-20180321164747-3a771d992973/go.mod h1:Dwedo/Wpr24TaqPxmxbtue+5NUziq4I4S80YR8gNf3Q= github.com/beorn7/perks v1.0.0/go.mod h1:KWe93zE9D1o94FZ5RNwFwVgaQK1VOXiVxmqh+CedLV8= github.com/beorn7/perks v1.0.1 h1:VlbKKnNfV8bJzeqoa4cOKqO6bYr3WgKZxO8Z16+hsOM= github.com/beorn7/perks v1.0.1/go.mod h1:G2ZrVWU2WbWT9wwq4/hrbKbnv/1ERSJQ0ibhJ6rlkpw= github.com/census-instrumentation/opencensus-proto v0.2.1/go.mod h1:f6KPmirojxKA12rnyqOA5BBL4O983OfeGPqjHWSTneU= github.com/census-instrumentation/opencensus-proto v0.4.1 h1:iKLQ0xPNFxR/2hzXZMrBo8f1j86j5WHzznCCQxV/b8g= -github.com/cespare/xxhash v1.1.0/go.mod h1:XrSqR1VqqWfGrhpAt58auRo0WTKS1nRRg3ghfAqPWnc= github.com/cespare/xxhash/v2 v2.1.1/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XLFGgcrjCOs= github.com/cespare/xxhash/v2 v2.1.2/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XLFGgcrjCOs= github.com/cespare/xxhash/v2 v2.2.0 h1:DC2CZ1Ep5Y4k3ZQ899DldepgrayRUGE6BBZ/cd9Cj44= @@ -91,46 +71,29 @@ github.com/chzyer/readline v0.0.0-20180603132655-2972be24d48e/go.mod h1:nSuG5e5P github.com/chzyer/test v0.0.0-20180213035817-a1ea475d72b1/go.mod h1:Q3SI9o4m/ZMnBNeIyt5eFwwo7qiLfzFZmjNmxjkiQlU= github.com/client9/misspell v0.3.4/go.mod h1:qj6jICC3Q7zFZvVWo7KLAzC3yx5G7kyvSDkc90ppPyw= github.com/cncf/udpa/go v0.0.0-20191209042840-269d4d468f6f/go.mod h1:M8M6+tZqaGXZJjfX53e64911xZQV5JYwmTeXPW+k8Sc= -github.com/cncf/udpa/go v0.0.0-20200629203442-efcf912fb354/go.mod h1:WmhPx2Nbnhtbo57+VJT5O0JRkEi1Wbu0z5j0R8u5Hbk= github.com/cncf/udpa/go v0.0.0-20201120205902-5459f2c99403/go.mod h1:WmhPx2Nbnhtbo57+VJT5O0JRkEi1Wbu0z5j0R8u5Hbk= github.com/cncf/udpa/go v0.0.0-20220112060539-c52dc94e7fbe h1:QQ3GSy+MqSHxm/d8nCtnAiZdYFd45cYZPs8vOOIYKfk= -github.com/cncf/xds/go v0.0.0-20210312221358-fbca930ec8ed/go.mod h1:eXthEFrGJvWHgFFCl3hGmgk+/aYT6PnTQLykKQRLhEs= github.com/cncf/xds/go v0.0.0-20230105202645-06c439db220b h1:ACGZRIr7HsgBKHsueQ1yM4WaVaXh21ynwqsF8M8tXhA= -github.com/cockroachdb/apd v1.1.0/go.mod h1:8Sl8LxpKi29FqWXR16WEFZRNSz3SoPzUzeMeY4+DwBQ= github.com/coreos/go-semver v0.3.0 h1:wkHLiw0WNATZnSG7epLsujiMCgPAc9xhjJ4tgnAxmfM= github.com/coreos/go-semver v0.3.0/go.mod h1:nnelYz7RCh+5ahJtPPxZlU+153eP4D4r3EedlOD2RNk= github.com/coreos/go-systemd/v22 v22.3.2 h1:D9/bQk5vlXQFZ6Kwuu6zaiXJ9oTPe68++AzAJc1DzSI= github.com/coreos/go-systemd/v22 v22.3.2/go.mod h1:Y58oyj3AT4RCenI/lSvhwexgC+NSVTIJ3seZv2GcEnc= github.com/cpuguy83/go-md2man/v2 v2.0.2/go.mod h1:tgQtvFlXSQOSOSIRvRPT7W67SCa46tRHOmNcaadrF8o= -github.com/creack/pty v1.1.9/go.mod h1:oKZEueFk5CKHvIhNR5MUki03XCEU+Q6VDXinZuGJ33E= github.com/davecgh/go-spew v1.1.0/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= github.com/davecgh/go-spew v1.1.1 h1:vj9j/u1bqnvCEfJOwUhtlOARqs3+rkHYY13jYWTU97c= github.com/davecgh/go-spew v1.1.1/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= -github.com/dgrijalva/jwt-go v0.0.0-20160705203006-01aeca54ebda/go.mod h1:E3ru+11k8xSBh+hMPgOLZmtrrCbhqsmaPHjLKYnJCaQ= -github.com/docker/spdystream v0.0.0-20160310174837-449fdfce4d96/go.mod h1:Qh8CwZgvJUkLughtfhJv5dyTYa91l1fOUCrgjqmcifM= github.com/dustin/go-humanize v1.0.0/go.mod h1:HtrtbFcZ19U5GC7JDqmcUSB87Iq5E25KnS6fMYU6eOk= -github.com/dustinkirkland/golang-petname v0.0.0-20191129215211-8e5a1ed0cff0/go.mod h1:V+Qd57rJe8gd4eiGzZyg4h54VLHmYVVw54iMnlAMrF8= -github.com/elazarl/goproxy v0.0.0-20170405201442-c4fc26588b6e/go.mod h1:/Zj4wYkgs4iZTTu3o/KG3Itv/qCCa8VVMlb3i9OVuzc= github.com/envoyproxy/go-control-plane v0.9.0/go.mod h1:YTl/9mNaCwkRvm6d1a2C3ymFceY/DCBVvsKhRF0iEA4= github.com/envoyproxy/go-control-plane v0.9.1-0.20191026205805-5f8ba28d4473/go.mod h1:YTl/9mNaCwkRvm6d1a2C3ymFceY/DCBVvsKhRF0iEA4= github.com/envoyproxy/go-control-plane v0.9.4/go.mod h1:6rpuAdCZL397s3pYoYcLgu1mIlRU8Am5FuJP05cCM98= -github.com/envoyproxy/go-control-plane v0.9.7/go.mod h1:cwu0lG7PUMfa9snN8LXBig5ynNVH9qI8YYLbd1fK2po= -github.com/envoyproxy/go-control-plane v0.9.9-0.20201210154907-fd9021fe5dad/go.mod h1:cXg6YxExXjJnVBQHBLXeUAgxn2UodCpnH306RInaBQk= github.com/envoyproxy/go-control-plane v0.9.9-0.20210217033140-668b12f5399d/go.mod h1:cXg6YxExXjJnVBQHBLXeUAgxn2UodCpnH306RInaBQk= -github.com/envoyproxy/go-control-plane v0.9.9-0.20210512163311-63b5d3c536b0/go.mod h1:hliV/p42l8fGbc6Y9bQ70uLwIvmJyVE5k4iMKlh8wCQ= github.com/envoyproxy/go-control-plane v0.10.3 h1:xdCVXxEe0Y3FQith+0cj2irwZudqGYvecuLB1HtdexY= github.com/envoyproxy/protoc-gen-validate v0.1.0/go.mod h1:iSmxcyjqTsJpI2R4NaDN7+kN2VEUnK/pcBlmesArF7c= github.com/envoyproxy/protoc-gen-validate v0.9.1 h1:PS7VIOgmSVhWUEeZwTe7z7zouA22Cr590PzXKbZHOVY= github.com/estuary/flow v0.1.9-0.20230303181027-f65a9d7f1a89 h1:qf9Pciyb/k63u9aZLmeBqp6zBP/BZdRQ1fYMeu2IaTs= github.com/estuary/flow v0.1.9-0.20230303181027-f65a9d7f1a89/go.mod h1:BMnABnmpvpgAVWt8lNlw7LBnwuyLmTd0asVzUFLsKi8= -github.com/evanphx/json-patch v4.2.0+incompatible/go.mod h1:50XU6AFN0ol/bzJsmQLiYLvXMP4fmwYFNcr97nuDLSk= github.com/evanphx/json-patch/v5 v5.6.0 h1:b91NhWfaz02IuVxO9faSllyAtNXHMPkC5J8sJCLunww= github.com/evanphx/json-patch/v5 v5.6.0/go.mod h1:G79N1coSVB93tBe7j6PhzjmR3/2VvlbKOFpnXhI9Bw4= -github.com/facebookgo/ensure v0.0.0-20200202191622-63f1cf65ac4c/go.mod h1:Yg+htXGokKKdzcwhuNDwVvN+uBxDGXJ7G/VN1d8fa64= -github.com/facebookgo/stack v0.0.0-20160209184415-751773369052/go.mod h1:UbMTZqLaRiH3MsBH8va0n7s1pQYcu3uTb8G4tygF4Zg= -github.com/facebookgo/subset v0.0.0-20200203212716-c811ad88dec4/go.mod h1:5tD+neXqOorC30/tWg0LCSkrqj/AR6gu8yY8/fpw1q0= -github.com/form3tech-oss/jwt-go v3.2.2+incompatible/go.mod h1:pbq4aXjuKjdthFRnoDwaVPLA+WlJuPGy+QneDUgJi2k= -github.com/fsnotify/fsnotify v1.4.7/go.mod h1:jwhsz4b93w/PPRr/qN1Yymfu8t87LnFCMoQvtojpjFo= github.com/ghodss/yaml v1.0.0/go.mod h1:4dBDuWmgqj2HViK6kFavaiC9ZROes6MMH2rRYeMEF04= github.com/go-gl/glfw v0.0.0-20190409004039-e6da0acd62b1/go.mod h1:vR7hzQXu2zJy9AVAgeJqvqgH9Q5CA+iKCZ2gyEVpxRU= github.com/go-gl/glfw/v3.3/glfw v0.0.0-20191125211704-12ad95a8df72/go.mod h1:tQ2UAYgL5IevRw8kRxooKSPJfGvJ9fJQFa0TUsXzTg8= @@ -147,16 +110,13 @@ github.com/go-stack/stack v1.8.0/go.mod h1:v0f6uXyyMGvRgIKkXu+yp6POWl0qKG85gN/me github.com/godbus/dbus/v5 v5.0.4/go.mod h1:xhWf0FNVPg57R7Z0UbKHbJfkEywrmjJnf7w5xrFpKfA= github.com/gogo/gateway v1.1.0 h1:u0SuhL9+Il+UbjM9VIE3ntfRujKbvVpFvNB4HbjeVQ0= github.com/gogo/gateway v1.1.0/go.mod h1:S7rR8FRQyG3QFESeSv4l2WnsyzlCLG0CzBbUUo/mbic= -github.com/gogo/protobuf v0.0.0-20171007142547-342cbe0a0415/go.mod h1:r8qH/GZQm5c6nD/R0oafs1akxWv10x8SbQlK7atdtwQ= github.com/gogo/protobuf v1.1.1/go.mod h1:r8qH/GZQm5c6nD/R0oafs1akxWv10x8SbQlK7atdtwQ= github.com/gogo/protobuf v1.2.0/go.mod h1:r8qH/GZQm5c6nD/R0oafs1akxWv10x8SbQlK7atdtwQ= -github.com/gogo/protobuf v1.2.1/go.mod h1:hp+jE20tsWTFYpLwKvXlhS1hjn+gTNwPg2I6zVXpSg4= github.com/gogo/protobuf v1.3.2 h1:Ov1cvc58UF3b5XjBnZv7+opcTcQFZebYjWzi34vdm4Q= github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69NZV8Q= -github.com/golang-jwt/jwt/v4 v4.4.1 h1:pC5DB52sCeK48Wlb9oPcdhnjkz1TKt1D/P7WKJ0kUcQ= -github.com/golang-jwt/jwt/v4 v4.4.1/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0= +github.com/golang-jwt/jwt/v4 v4.5.0 h1:7cYmW1XlMY7h7ii7UhUyChSgS5wUJEnm9uZVTGqOWzg= +github.com/golang-jwt/jwt/v4 v4.5.0/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0= github.com/golang/glog v0.0.0-20160126235308-23def4e6c14b/go.mod h1:SBH7ygxi8pfUlaOkMMuAQtPIUF8ecWP5IEl/CR7VP2Q= -github.com/golang/groupcache v0.0.0-20160516000752-02826c3e7903/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/groupcache v0.0.0-20190702054246-869f871628b6/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/groupcache v0.0.0-20191227052852-215e87163ea7/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc= github.com/golang/groupcache v0.0.0-20200121045136-8c9f03a8e57e h1:1r7pUrabqp18hOBcwBwiTsbnFeTZHV9eER/QT5JVZxY= @@ -168,8 +128,6 @@ github.com/golang/mock v1.4.0/go.mod h1:UOMv5ysSaYNkG+OFQykRIcU/QvvxJf3p21QfJ2Bt github.com/golang/mock v1.4.1/go.mod h1:UOMv5ysSaYNkG+OFQykRIcU/QvvxJf3p21QfJ2Bt3cw= github.com/golang/mock v1.4.3/go.mod h1:UOMv5ysSaYNkG+OFQykRIcU/QvvxJf3p21QfJ2Bt3cw= github.com/golang/mock v1.4.4/go.mod h1:l3mdAwkq5BuhzHwde/uurv3sEJeZMXNpwsxVWU71h+4= -github.com/golang/mock v1.5.0/go.mod h1:CWnOUgYIOo4TcNZ0wHX3YZCqsaM1I1Jvs6v3mP3KVu8= -github.com/golang/mock v1.6.0/go.mod h1:p6yTPP+5HYm5mzsMV8JkE6ZKdX+/wYM6Hr+LicevLPs= github.com/golang/protobuf v1.2.0/go.mod h1:6lQm79b+lXiMfvg/cZm0SGofjICqVBUtrP5yJMmIC1U= github.com/golang/protobuf v1.3.1/go.mod h1:6lQm79b+lXiMfvg/cZm0SGofjICqVBUtrP5yJMmIC1U= github.com/golang/protobuf v1.3.2/go.mod h1:6lQm79b+lXiMfvg/cZm0SGofjICqVBUtrP5yJMmIC1U= @@ -185,13 +143,10 @@ github.com/golang/protobuf v1.4.1/go.mod h1:U8fpvMrcmy5pZrNK1lt4xCsGvpyWQ/VVv6QD github.com/golang/protobuf v1.4.2/go.mod h1:oDoupMAO8OvCJWAcko0GGGIgR6R6ocIYbsSw735rRwI= github.com/golang/protobuf v1.4.3/go.mod h1:oDoupMAO8OvCJWAcko0GGGIgR6R6ocIYbsSw735rRwI= github.com/golang/protobuf v1.5.0/go.mod h1:FsONVRAS9T7sI+LIUmWTfcYkHO4aIWwzhcaSAoJOfIk= -github.com/golang/protobuf v1.5.1/go.mod h1:DopwsBzvsk0Fs44TXzsVbJyPhcCPeIwnvohx4u74HPM= github.com/golang/protobuf v1.5.2 h1:ROPKBNFfQgOUMifHyP+KYbvpjbdoFNs+aK7DXlji0Tw= github.com/golang/protobuf v1.5.2/go.mod h1:XVQd3VNwM+JqD3oG2Ue2ip4fOMUkwXdXDdiuN0vRsmY= -github.com/golang/snappy v0.0.3/go.mod h1:/XxbfmMg8lxefKM7IXC3fBNl/7bRcc72aCRzEWrmP2Q= github.com/golang/snappy v0.0.4 h1:yAGX7huGHXlcLOEtBnF4w7FQwA26wojNCwOYAEhLjQM= github.com/golang/snappy v0.0.4/go.mod h1:/XxbfmMg8lxefKM7IXC3fBNl/7bRcc72aCRzEWrmP2Q= -github.com/google/btree v0.0.0-20160524151835-7d79101e329e/go.mod h1:lNA+9X1NB3Zf8V7Ke586lFgjr2dZNuvo3lPJSGZ5JPQ= github.com/google/btree v0.0.0-20180813153112-4030bb1f1f0c/go.mod h1:lNA+9X1NB3Zf8V7Ke586lFgjr2dZNuvo3lPJSGZ5JPQ= github.com/google/btree v1.0.0/go.mod h1:lNA+9X1NB3Zf8V7Ke586lFgjr2dZNuvo3lPJSGZ5JPQ= github.com/google/go-cmp v0.2.0/go.mod h1:oXzfMopK8JAjlY9xF4vHSVASa0yLyX7SntLO5aqRK0M= @@ -201,17 +156,12 @@ github.com/google/go-cmp v0.4.0/go.mod h1:v8dTdLbMG2kIc/vJvl+f65V22dbkXbowE6jgT/ github.com/google/go-cmp v0.4.1/go.mod h1:v8dTdLbMG2kIc/vJvl+f65V22dbkXbowE6jgT/gNBxE= github.com/google/go-cmp v0.5.0/go.mod h1:v8dTdLbMG2kIc/vJvl+f65V22dbkXbowE6jgT/gNBxE= github.com/google/go-cmp v0.5.1/go.mod h1:v8dTdLbMG2kIc/vJvl+f65V22dbkXbowE6jgT/gNBxE= -github.com/google/go-cmp v0.5.2/go.mod h1:v8dTdLbMG2kIc/vJvl+f65V22dbkXbowE6jgT/gNBxE= -github.com/google/go-cmp v0.5.3/go.mod h1:v8dTdLbMG2kIc/vJvl+f65V22dbkXbowE6jgT/gNBxE= github.com/google/go-cmp v0.5.4/go.mod h1:v8dTdLbMG2kIc/vJvl+f65V22dbkXbowE6jgT/gNBxE= github.com/google/go-cmp v0.5.5/go.mod h1:v8dTdLbMG2kIc/vJvl+f65V22dbkXbowE6jgT/gNBxE= -github.com/google/go-cmp v0.5.6/go.mod h1:v8dTdLbMG2kIc/vJvl+f65V22dbkXbowE6jgT/gNBxE= github.com/google/go-cmp v0.5.9 h1:O2Tfq5qg4qc4AmwVlvv0oLiVAGB7enBSJ2x2DqQFi38= github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg= github.com/google/martian v2.1.0+incompatible/go.mod h1:9I4somxYTbIHy5NJKHRl3wXiIaQGbYVAs8BPL6v8lEs= github.com/google/martian/v3 v3.0.0/go.mod h1:y5Zk1BBys9G+gd6Jrk0W3cC1+ELVxBWuIGO+w/tUAp0= -github.com/google/martian/v3 v3.1.0/go.mod h1:y5Zk1BBys9G+gd6Jrk0W3cC1+ELVxBWuIGO+w/tUAp0= -github.com/google/martian/v3 v3.2.1/go.mod h1:oBOf6HBosgwRXnUGWUB05QECsc6uvmMiJ3+6W4l/CUk= github.com/google/pprof v0.0.0-20181206194817-3ea8567a2e57/go.mod h1:zfwlbNMJ+OItoe0UupaVj+oy1omPYYDuagoSzA8v9mc= github.com/google/pprof v0.0.0-20190515194954-54271f7e092f/go.mod h1:zfwlbNMJ+OItoe0UupaVj+oy1omPYYDuagoSzA8v9mc= github.com/google/pprof v0.0.0-20191218002539-d4f498aebedc/go.mod h1:ZgVRPoUq/hfqzAqh7sHMqb3I9Rq5C59dIz2SbBwJ4eM= @@ -219,28 +169,14 @@ github.com/google/pprof v0.0.0-20200212024743-f11f1df84d12/go.mod h1:ZgVRPoUq/hf github.com/google/pprof v0.0.0-20200229191704-1ebb73c60ed3/go.mod h1:ZgVRPoUq/hfqzAqh7sHMqb3I9Rq5C59dIz2SbBwJ4eM= github.com/google/pprof v0.0.0-20200430221834-fc25d7d30c6d/go.mod h1:ZgVRPoUq/hfqzAqh7sHMqb3I9Rq5C59dIz2SbBwJ4eM= github.com/google/pprof v0.0.0-20200708004538-1a94d8640e99/go.mod h1:ZgVRPoUq/hfqzAqh7sHMqb3I9Rq5C59dIz2SbBwJ4eM= -github.com/google/pprof v0.0.0-20201023163331-3e6fc7fc9c4c/go.mod h1:kpwsk12EmLew5upagYY7GY0pfYCcupk39gWOCRROcvE= -github.com/google/pprof v0.0.0-20201203190320-1bf35d6f28c2/go.mod h1:kpwsk12EmLew5upagYY7GY0pfYCcupk39gWOCRROcvE= -github.com/google/pprof v0.0.0-20210122040257-d980be63207e/go.mod h1:kpwsk12EmLew5upagYY7GY0pfYCcupk39gWOCRROcvE= -github.com/google/pprof v0.0.0-20210226084205-cbba55b83ad5/go.mod h1:kpwsk12EmLew5upagYY7GY0pfYCcupk39gWOCRROcvE= -github.com/google/pprof v0.0.0-20210601050228-01bbb1931b22/go.mod h1:kpwsk12EmLew5upagYY7GY0pfYCcupk39gWOCRROcvE= -github.com/google/pprof v0.0.0-20210609004039-a478d1d731e9/go.mod h1:kpwsk12EmLew5upagYY7GY0pfYCcupk39gWOCRROcvE= -github.com/google/pprof v0.0.0-20210720184732-4bb14d4b1be1/go.mod h1:kpwsk12EmLew5upagYY7GY0pfYCcupk39gWOCRROcvE= github.com/google/renameio v0.1.0/go.mod h1:KWCgfxg9yswjAJkECMjeO8J8rahYeXnNhOm40UhjYkI= -github.com/google/uuid v1.1.1/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/google/uuid v1.1.2/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= -github.com/google/uuid v1.2.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/google/uuid v1.3.0 h1:t6JiXgmwXMjEs8VusXIJk2BXHsn+wx8BZdTaoZ5fu7I= github.com/google/uuid v1.3.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= github.com/googleapis/gax-go/v2 v2.0.4/go.mod h1:0Wqv26UfaUD9n4G6kQubkQ+KchISgw+vpHVxEJEs9eg= github.com/googleapis/gax-go/v2 v2.0.5/go.mod h1:DWXyrwAJ9X0FpwwEdw+IPEYBICEFu5mhpdKc/us6bOk= -github.com/googleapis/gax-go/v2 v2.1.0/go.mod h1:Q3nei7sK6ybPYH7twZdmQpAd1MKb7pfu6SK+H1/DsU0= github.com/googleapis/gax-go/v2 v2.1.1 h1:dp3bWCh+PPO1zjRRiCSczJav13sBvG4UhNyVTa1KqdU= -github.com/googleapis/gnostic v0.0.0-20170729233727-0c5108395e2d/go.mod h1:sJBsCZ4ayReDTBIg8b9dl28c5xFWyhBTVRp3pOg5EKY= -github.com/gophercloud/gophercloud v0.0.0-20190126172459-c818fa66e4c8/go.mod h1:3WdhXV3rUYy9p6AUW8d94kr+HS62Y4VL9mBnFxsD8q4= github.com/gorilla/schema v1.2.0 h1:YufUaxZYCKGFuAq3c96BOhjgd5nmXiOY9NGzF247Tsc= -github.com/gorilla/schema v1.2.0/go.mod h1:kgLaKoK1FELgZqMAVxx/5cbj0kT+57qxUrAlIO2eleU= -github.com/gregjones/httpcache v0.0.0-20170728041850-787624de3eb7/go.mod h1:FecbI9+v66THATjSRHfNgh1IVFe/9kFxbXtjV0ctIMA= github.com/grpc-ecosystem/go-grpc-prometheus v1.2.0 h1:Ovs26xHkKqVztRpIrF/92BcuyuQ/YW4NSIpoGtfXNho= github.com/grpc-ecosystem/go-grpc-prometheus v1.2.0/go.mod h1:8NvIoxWQoOIhqOTXgfV/d3M/q6VIi02HzZEHgUlZvzk= github.com/grpc-ecosystem/grpc-gateway v1.8.5/go.mod h1:vNeuVxBJEsws4ogUvrchl83t/GYV9WGTSLVdBhOQFDY= @@ -252,23 +188,13 @@ github.com/hashicorp/golang-lru v0.5.4 h1:YDjusn29QI/Das2iO9M0BHnIbxPeyuCHsjMW+l github.com/hashicorp/golang-lru v0.5.4/go.mod h1:iADmTwqILo4mZ8BN3D2Q6+9jd8WM5uGBxy+E8yxSoD4= github.com/hashicorp/golang-lru/v2 v2.0.1 h1:5pv5N1lT1fjLg2VQ5KWc7kmucp2x/kvFOnxuVTqZ6x4= github.com/hashicorp/golang-lru/v2 v2.0.1/go.mod h1:QeFd9opnmA6QUJc5vARoKUSoFhyfM2/ZepoAG6RGpeM= -github.com/hpcloud/tail v1.0.0/go.mod h1:ab1qPbhIpdTxEkNHXyeSf5vhxWSCs/tWer42PpOxQnU= github.com/ianlancetaylor/demangle v0.0.0-20181102032728-5e5cf60278f6/go.mod h1:aSSvb/t6k1mPoxDqO4vJh6VOCGPwU4O0C2/Eqndh1Sc= -github.com/ianlancetaylor/demangle v0.0.0-20200824232613-28f6c0f3b639/go.mod h1:aSSvb/t6k1mPoxDqO4vJh6VOCGPwU4O0C2/Eqndh1Sc= -github.com/imdario/mergo v0.3.5/go.mod h1:2EnlNZ0deacrJVfApfmtdGgDfMuh/nq6Ok1EcJh5FfA= github.com/inconshreveable/mousetrap v1.0.1 h1:U3uMjPSQEBMNp1lFxmllqCPM6P5u/Xq7Pgzkat/bFNc= github.com/inconshreveable/mousetrap v1.0.1/go.mod h1:vpF70FUmC8bwa3OWnCshd2FqLfsEA9PFc4w1p2J65bw= github.com/jamiealquiza/envy v1.1.0 h1:Nwh4wqTZ28gDA8zB+wFkhnUpz3CEcO12zotjeqqRoKE= github.com/jamiealquiza/envy v1.1.0/go.mod h1:MP36BriGCLwEHhi1OU8E9569JNZrjWfCvzG7RsPnHus= github.com/jessevdk/go-flags v1.4.0/go.mod h1:4FA24M0QyGHXBuZZK/XkWh8h0e1EYbRYJSGM75WSRxI= -github.com/jessevdk/go-flags v1.4.1-0.20181221193153-c0795c8afcf4/go.mod h1:4FA24M0QyGHXBuZZK/XkWh8h0e1EYbRYJSGM75WSRxI= -github.com/jessevdk/go-flags v1.5.0/go.mod h1:Fw0T6WPc1dYxT4mKEZRfG5kJhaTDP9pj1c2EWnYs/m4= -github.com/jgraettinger/cockroach-encoding v1.1.0/go.mod h1:gcht+UqiTiDC6NEF7DLHUXQStSlWjogvhlkAttXVtmE= -github.com/jgraettinger/gorocksdb v0.0.0-20210726190246-aede64bf0e7b/go.mod h1:pF+aV5pQJlgi8ZKGnuUhEzLF1Z74ONYqrn0O2UXLjeQ= -github.com/jgraettinger/urkel v0.1.2/go.mod h1:PQ2/GQdeAtepWvfbEczTdzYM+bmB0qgB0L23qeUaKbA= github.com/jmespath/go-jmespath v0.4.0 h1:BEgLn5cpjn8UN1mAw4NjwDrS35OdebyEtFe+9YPoQUg= -github.com/jmespath/go-jmespath v0.4.0/go.mod h1:T8mJZnbsbmF+m6zOOFylbeCJqk5+pHWvzYPziyZiYoo= -github.com/jmespath/go-jmespath/internal/testify v1.5.1/go.mod h1:L3OGu8Wl2/fWfCI6z80xFu9LTZmf1ZRjMHUOPmWr69U= github.com/jpillora/backoff v1.0.0/go.mod h1:J/6gKK9jxlEcS3zixgDgUAsiuZ7yrSoa/FX5e0EB2j4= github.com/json-iterator/go v1.1.6/go.mod h1:+SdeFBvtyEkXs7REEP0seUULqWtbJapLOCVDaaPEHmU= github.com/json-iterator/go v1.1.10/go.mod h1:KdQUCv79m/52Kvf8AW2vK1V8akMuk1QjK/uOdHXbAo4= @@ -278,10 +204,8 @@ github.com/jstemmer/go-junit-report v0.0.0-20190106144839-af01ea7f8024/go.mod h1 github.com/jstemmer/go-junit-report v0.9.1/go.mod h1:Brl9GWCQeLvo8nXZwPNNblvFj/XSXhF0NWZEnDohbsk= github.com/julienschmidt/httprouter v1.2.0/go.mod h1:SYymIcj16QtmaHHD7aYtjjsJG7VTCxuUUipMqKk8s4w= github.com/julienschmidt/httprouter v1.3.0/go.mod h1:JR6WtHb+2LUe8TCKY3cZOxFyyO8IZAc4RVcycCCAKdM= -github.com/kisielk/errcheck v1.1.0/go.mod h1:EZBBE59ingxPouuu3KfxchcWSUPOHkagtvWXihfKN4Q= github.com/kisielk/errcheck v1.5.0/go.mod h1:pFxgyoBC7bSaBwPgfKdkLd5X25qrDl4LWUI2bnpBCr8= github.com/kisielk/gotool v1.0.0/go.mod h1:XhKaO+MFFWcvkIS/tQcRk01m1F5IRFswLeQ+oQHNcck= -github.com/klauspost/compress v1.13.5/go.mod h1:/3/Vjq9QcHkK5uEr5lBEmyoZ1iFhe47etQ6QUkpK6sk= github.com/klauspost/compress v1.16.3 h1:XuJt9zzcnaz6a16/OU53ZjWp/v7/42WcR5t2a0PcNQY= github.com/klauspost/compress v1.16.3/go.mod h1:ntbaceVETuRiXiv4DpjP66DpAtAGkEQskQzEyD//IeE= github.com/konsorten/go-windows-terminal-sequences v1.0.1/go.mod h1:T0+1ngSBFLxvqU3pZ+m/2kptfBszLMUkC4ZK/EgS/cQ= @@ -289,19 +213,12 @@ github.com/konsorten/go-windows-terminal-sequences v1.0.3/go.mod h1:T0+1ngSBFLxv github.com/kr/fs v0.1.0/go.mod h1:FFnZGqtBN9Gxj7eW1uZ42v5BccTP0vu6NEaFoC2HwRg= github.com/kr/logfmt v0.0.0-20140226030751-b84e30acd515/go.mod h1:+0opPa2QZZtGFBFZlji/RkVcI2GknAs/DXo4wKdlNEc= github.com/kr/pretty v0.1.0/go.mod h1:dAy3ld7l9f0ibDNOQOHHMYYIIbhfbHSm3C4ZsoJORNo= -github.com/kr/pretty v0.2.1/go.mod h1:ipq/a2n7PKx3OHsz4KJII5eveXtPO4qwEXGdVfWzfnI= github.com/kr/pretty v0.3.0 h1:WgNl7dwNpEZ6jJ9k1snq4pZsg7DOEN8hP9Xw0Tsjwk0= -github.com/kr/pretty v0.3.0/go.mod h1:640gp4NfQd8pI5XOwp5fnNeVWj67G7CFk/SaSQn7NBk= github.com/kr/pty v1.1.1/go.mod h1:pFQYn66WHrOpPYNljwOMqo10TkYh1fy3cYio2l3bCsQ= github.com/kr/text v0.1.0/go.mod h1:4Jbv+DJW3UT/LiOwJeYQe1efqtUx/iVham/4vfdArNI= github.com/kr/text v0.2.0 h1:5Nx0Ya0ZqY2ygV366QzturHI13Jq95ApcVaJBhpS+AY= -github.com/kr/text v0.2.0/go.mod h1:eLer722TekiGuMkidMxC/pM04lWEeraHUUmBw8l2grE= -github.com/lib/pq v1.10.2/go.mod h1:AlVN5x4E4T544tWzH6hKfbfQvm3HdbOxrmggDNAPY9o= +github.com/kylelemons/godebug v1.1.0 h1:RPNrshWIDI6G2gRW9EHilWtl7Z6Sb1BR0xunSBf0SNc= github.com/mattn/go-ieproxy v0.0.1 h1:qiyop7gCflfhwCzGyeT0gro3sF9AIg9HU98JORTkqfI= -github.com/mattn/go-ieproxy v0.0.1/go.mod h1:pYabZ6IHcRpFh7vIaLfK7rdcWgFEb3SFJ6/gNWuh88E= -github.com/mattn/go-runewidth v0.0.9/go.mod h1:H031xJmbD/WCDINGzjvQ9THkh0rPKHF+m2gUSrubnMI= -github.com/mattn/go-runewidth v0.0.13/go.mod h1:Jdepj2loyihRzMpdS35Xk/zdY8IAYHsh153qUoGf23w= -github.com/mattn/go-sqlite3 v2.0.3+incompatible/go.mod h1:FPy6KqzDD04eiIsT53CuJW3U88zkxoIYsOqkbpncsNc= github.com/matttproud/golang_protobuf_extensions v1.0.1/go.mod h1:D8He9yQNgCq6Z5Ld7szi9bcBfOoFv/3dc6xSMkL2PC0= github.com/matttproud/golang_protobuf_extensions v1.0.2-0.20181231171920-c182affec369 h1:I0XW9+e1XWDxdcEniV4rQAIOPUGDq67JSCiRCgGCZLI= github.com/matttproud/golang_protobuf_extensions v1.0.2-0.20181231171920-c182affec369/go.mod h1:BSXmuO+STAnVfrANrmjBb36TMTDstsz7MSK+HVaYKv4= @@ -312,13 +229,7 @@ github.com/modern-go/reflect2 v1.0.1/go.mod h1:bx2lNnkwVCuqBIxFjflWJWanXIb3Rllmb github.com/modern-go/reflect2 v1.0.2/go.mod h1:yWuevngMOJpCy52FWWMvUC8ws7m/LJsjYzDa0/r8luk= github.com/mwitkow/go-conntrack v0.0.0-20161129095857-cc309e4a2223/go.mod h1:qRWi+5nqEBWmkhHvq77mSJWrCKwh8bxhgT7d/eI7P4U= github.com/mwitkow/go-conntrack v0.0.0-20190716064945-2f068394615f/go.mod h1:qRWi+5nqEBWmkhHvq77mSJWrCKwh8bxhgT7d/eI7P4U= -github.com/mxk/go-flowrate v0.0.0-20140419014527-cca7078d478f/go.mod h1:ZdcZmHo+o7JKHSa8/e818NopupXU1YMK5fe1lsApnBw= -github.com/olekukonko/tablewriter v0.0.5/go.mod h1:hPp6KlRPjbx+hW8ykQs1w3UBbZlj6HuIJcUGPhkA7kY= -github.com/onsi/ginkgo v1.6.0/go.mod h1:lLunBs/Ym6LB5Z9jYTR76FiuTmxDTDusOGeTQH+WWjE= -github.com/onsi/ginkgo v1.8.0/go.mod h1:lLunBs/Ym6LB5Z9jYTR76FiuTmxDTDusOGeTQH+WWjE= -github.com/onsi/gomega v1.5.0/go.mod h1:ex+gbHU/CVuBBDIJjb2X0qEXbFg53c61hWP/1CpauHY= -github.com/peterbourgon/diskv v2.0.1+incompatible/go.mod h1:uqqh8zWWbv1HBMNONnaR/tNboyR3/BZd58JJSHlUSCU= -github.com/pkg/diff v0.0.0-20210226163009-20ebb0f2a09e/go.mod h1:pJLUxLENpZxwdsKMEsNbx1VGcRFpLqf3715MtcvvzbA= +github.com/pkg/browser v0.0.0-20210911075715-681adbf594b8 h1:KoWmjvw+nsYOo29YJK9vDA65RGE3NrOnUtO7a+RF9HU= github.com/pkg/errors v0.8.0/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pkg/errors v0.8.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0= github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4= @@ -343,7 +254,6 @@ github.com/prometheus/client_model v0.3.0/go.mod h1:LDGWKZIo7rky3hgvBe+caln+Dr3d github.com/prometheus/common v0.4.1/go.mod h1:TNfzLD0ON7rHzMJeJkieUDPYmFC7Snx/y86RQel1bk4= github.com/prometheus/common v0.10.0/go.mod h1:Tlit/dnDKsSWFlCLTWaA1cyBgKHSMdTB80sz/V91rCo= github.com/prometheus/common v0.26.0/go.mod h1:M7rCNAaPfAosfx8veZJCuw84e35h3Cfd9VFqTh1DIvc= -github.com/prometheus/common v0.30.0/go.mod h1:vu+V0TpY+O6vW9J44gczi3Ap/oXXR10b+M/gUGO4Hls= github.com/prometheus/common v0.32.1/go.mod h1:vu+V0TpY+O6vW9J44gczi3Ap/oXXR10b+M/gUGO4Hls= github.com/prometheus/common v0.37.0 h1:ccBbHCgIiT9uSoFY0vX8H3zsNR5eLt17/RQLUvn8pXE= github.com/prometheus/common v0.37.0/go.mod h1:phzohg0JFMnBEFGxTDbfu3QyL5GI8gTQJFhYO5B3mfA= @@ -354,30 +264,21 @@ github.com/prometheus/procfs v0.6.0/go.mod h1:cz+aTbrPOrUb4q7XlbU9ygM+/jj0fzG6c1 github.com/prometheus/procfs v0.7.3/go.mod h1:cz+aTbrPOrUb4q7XlbU9ygM+/jj0fzG6c1xBZuNvfVA= github.com/prometheus/procfs v0.8.0 h1:ODq8ZFEaYeCaZOJlZZdJA2AbQR98dSHSM1KW/You5mo= github.com/prometheus/procfs v0.8.0/go.mod h1:z7EfXMXOkbkqb9IINtpCn86r/to3BnA0uaxHdg830/4= -github.com/rivo/uniseg v0.2.0/go.mod h1:J6wj4VEh+S6ZtnVlnTBMWIodfgj8LQOQFoIToxlJtxc= github.com/rogpeppe/fastuuid v0.0.0-20150106093220-6724a57986af/go.mod h1:XWv6SoW27p1b0cqNHllgS5HIMJraePCO15w5zCzIWYg= github.com/rogpeppe/fastuuid v1.2.0/go.mod h1:jVj6XXZzXRy/MSR5jhDC/2q6DgLz+nrA6LYCDYWNEvQ= github.com/rogpeppe/go-internal v1.3.0/go.mod h1:M8bDsm7K2OlrFYOpmOWEs/qY81heoFRclV5y23lUDJ4= -github.com/rogpeppe/go-internal v1.6.1/go.mod h1:xXDCJY+GAPziupqXw64V24skbSoqbTEfhy4qGm1nDQc= github.com/rogpeppe/go-internal v1.8.0 h1:FCbCCtXNOY3UtUuHUYaghJg4y7Fd14rXifAYUAtL9R8= -github.com/rogpeppe/go-internal v1.8.0/go.mod h1:WmiCO8CzOY8rg0OYDC4/i/2WRWAB6poM+XZ2dLUbcbE= github.com/russross/blackfriday/v2 v2.1.0/go.mod h1:+Rmxgy9KzJVeS9/2gXHxylqXiyQDYRxCVz55jmeOWTM= -github.com/satori/go.uuid v1.2.0/go.mod h1:dA0hQrYB0VpLJoorglMZABFdXlWrHn1NEOzdhQKdks0= -github.com/sirupsen/logrus v1.0.5/go.mod h1:pMByvHTf9Beacp5x1UXfOR9xyW/9antXMhjMPG0dEzc= github.com/sirupsen/logrus v1.2.0/go.mod h1:LxeOpSwHxABJmUn/MG1IvRgCAasNZTLOkJPxbbu5VWo= github.com/sirupsen/logrus v1.4.2/go.mod h1:tLMulIdttU9McNUspp0xgXVQah82FyeX6MwdIuYE2rE= github.com/sirupsen/logrus v1.6.0/go.mod h1:7uNnSEd1DgxDLC74fIahvMZmmYsHGZGEOFrfsX/uA88= -github.com/sirupsen/logrus v1.8.1/go.mod h1:yWOB1SBYBC5VeMP7gHvWumXLIWorT60ONWic61uBYv0= github.com/sirupsen/logrus v1.9.0 h1:trlNQbNUG3OdDrDil03MCb1H2o9nJ1x4/5LYw7byDE0= github.com/sirupsen/logrus v1.9.0/go.mod h1:naHLuLoDiP4jHNo9R0sCBMtWGeIprob74mVsIT4qYEQ= github.com/soheilhy/cmux v0.1.5 h1:jjzc5WVemNEDTLwv9tlmemhC73tI08BNOIGwBOo10Js= -github.com/soheilhy/cmux v0.1.5/go.mod h1:T7TcVDs9LWfQgPlPsdngu6I6QIoyIFZDDC6sNE1GqG0= -github.com/spaolacci/murmur3 v0.0.0-20180118202830-f09979ecbc72/go.mod h1:JwIasOWyU6f++ZhiEuf87xNszmSA2myDM2Kzu9HwQUA= github.com/spf13/afero v1.6.0 h1:xoax2sJ2DT8S8xA2paPFjDCScCNeWsg75VG0DLRreiY= github.com/spf13/afero v1.6.0/go.mod h1:Ai8FlHk4v/PARR026UzYexafAt9roJ7LcLMAmO6Z93I= github.com/spf13/cobra v1.6.1 h1:o94oiPyS4KD1mPy2fmcYYHHfCxLqYjJOhGsCHFZtEzA= github.com/spf13/cobra v1.6.1/go.mod h1:IOw/AERYS7UzyrGinqmz6HLUo219MORXGxhbaJUqzrY= -github.com/spf13/pflag v1.0.3/go.mod h1:DYY7MBk1bdzusC3SYhjObp+wFpr4gzcvqqNjLnInEg4= github.com/spf13/pflag v1.0.5 h1:iy+VFUOCP1a+8yFto/drg2CJ5u0yRoB7fZw3DKv/JXA= github.com/spf13/pflag v1.0.5/go.mod h1:McXfInJRrz4CZXVZOBLb0bTZqETkiAhM9Iw0y3An2Bg= github.com/stretchr/objx v0.1.0/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME= @@ -386,9 +287,8 @@ github.com/stretchr/testify v1.2.2/go.mod h1:a8OnRcib4nhh0OaRAV+Yts87kKdq0PP7pXf github.com/stretchr/testify v1.3.0/go.mod h1:M5WIy9Dh21IEIfnGCwXGc5bZfKNJtfHm1UVUgZn+9EI= github.com/stretchr/testify v1.4.0/go.mod h1:j7eGeouHqKxXV5pUuKE4zz7dFj8WfuZ+81PSLYec5m4= github.com/stretchr/testify v1.5.1/go.mod h1:5W2xD1RspED5o8YsWQXVCued0rvSQ+mT+I5cxcmMvtA= -github.com/stretchr/testify v1.6.1/go.mod h1:6Fq8oRcR53rry900zMqJjRRixrwX3KX962/h/Wwjteg= -github.com/stretchr/testify v1.7.0 h1:nwc3DEeHmmLAfoZucVR881uASk0Mfjw8xYJ99tb5CcY= github.com/stretchr/testify v1.7.0/go.mod h1:6Fq8oRcR53rry900zMqJjRRixrwX3KX962/h/Wwjteg= +github.com/stretchr/testify v1.7.1 h1:5TQK59W5E3v0r2duFAb7P95B6hEeOyEnHRa8MjYSMTY= github.com/urfave/negroni v1.0.0 h1:kIimOitoypq34K7TG7DUaJ9kq/N4Ofuwi1sjz0KipXc= github.com/urfave/negroni v1.0.0/go.mod h1:Meg73S6kFm/4PpbYdq35yYWoCZ9mS/YSx+lKnmiohz4= github.com/valyala/bytebufferpool v1.0.0 h1:GqA5TC/0021Y/b9FG4Oi9Mr3q7XYx6KllzawFIhcdPw= @@ -400,45 +300,34 @@ github.com/yuin/goldmark v1.1.27/go.mod h1:3hX8gzYuyVAZsxl0MRgGTJEmQBFcNTphYh9de github.com/yuin/goldmark v1.1.32/go.mod h1:3hX8gzYuyVAZsxl0MRgGTJEmQBFcNTphYh9decYSb74= github.com/yuin/goldmark v1.2.1/go.mod h1:3hX8gzYuyVAZsxl0MRgGTJEmQBFcNTphYh9decYSb74= github.com/yuin/goldmark v1.3.5/go.mod h1:mwnBkeHKe2W/ZEtQ+71ViKU8L12m81fl3OWwC1Zlc8k= -go.etcd.io/etcd/api/v3 v3.5.0/go.mod h1:cbVKeC6lCfl7j/8jBhAK6aIYO9XOjdptoxU/nLQcPvs= go.etcd.io/etcd/api/v3 v3.5.4 h1:OHVyt3TopwtUQ2GKdd5wu3PmmipR4FTwCqoEjSyRdIc= go.etcd.io/etcd/api/v3 v3.5.4/go.mod h1:5GB2vv4A4AOn3yk7MftYGHkUfGtDHnEraIjym4dYz5A= -go.etcd.io/etcd/client/pkg/v3 v3.5.0/go.mod h1:IJHfcCEKxYu1Os13ZdwCwIUTUVGYTSAM3YSwc9/Ac1g= go.etcd.io/etcd/client/pkg/v3 v3.5.4 h1:lrneYvz923dvC14R54XcA7FXoZ3mlGZAgmwhfm7HqOg= go.etcd.io/etcd/client/pkg/v3 v3.5.4/go.mod h1:IJHfcCEKxYu1Os13ZdwCwIUTUVGYTSAM3YSwc9/Ac1g= -go.etcd.io/etcd/client/v3 v3.5.0/go.mod h1:AIKXXVX/DQXtfTEqBryiLTUXwON+GuvO6Z7lLS/oTh0= go.etcd.io/etcd/client/v3 v3.5.4 h1:p83BUL3tAYS0OT/r0qglgc3M1JjhM0diV8DSWAhVXv4= go.etcd.io/etcd/client/v3 v3.5.4/go.mod h1:ZaRkVgBZC+L+dLCjTcF1hRXpgZXQPOvnA/Ak/gq3kiY= -go.gazette.dev/core v0.89.1-0.20230106205712-a2d374cdd6be h1:U1VCL/6BkY8nfPOZCxIQhtCHCQJTDQeo+GcOfiLgOJ0= -go.gazette.dev/core v0.89.1-0.20230106205712-a2d374cdd6be/go.mod h1:c/5g5752X3AH+g1ItiQaq9x+gmCFRmdTCFSkp5hypVQ= +go.gazette.dev/core v0.89.1-0.20231012132739-dfed675b7fd1 h1:oRtrsDKboiCGeG0B6kJvaJ9KVAn7U797oGMGTqdHnwo= +go.gazette.dev/core v0.89.1-0.20231012132739-dfed675b7fd1/go.mod h1:/fdxqReMWKS26yROKEKDY8JlXjwz8TqeEZggpNwb5I0= go.opencensus.io v0.21.0/go.mod h1:mSImk1erAIZhrmZN+AvHh14ztQfjbGwt4TtuofqLduU= go.opencensus.io v0.22.0/go.mod h1:+kGneAE2xo2IficOXnaByMWTGM9T73dGwxeWcUqIpI8= go.opencensus.io v0.22.2/go.mod h1:yxeiOL68Rb0Xd1ddK5vPZ/oVn4vY4Ynel7k9FzqtOIw= go.opencensus.io v0.22.3/go.mod h1:yxeiOL68Rb0Xd1ddK5vPZ/oVn4vY4Ynel7k9FzqtOIw= go.opencensus.io v0.22.4/go.mod h1:yxeiOL68Rb0Xd1ddK5vPZ/oVn4vY4Ynel7k9FzqtOIw= -go.opencensus.io v0.22.5/go.mod h1:5pWMHQbX5EPX2/62yrJeAkowc+lfs/XD7Uxpq3pI6kk= go.opencensus.io v0.23.0 h1:gqCw0LfLxScz8irSi8exQc7fyQ0fKQU/qnC/X8+V/1M= -go.opencensus.io v0.23.0/go.mod h1:XItmlyltB5F7CS4xOC1DcqMoFqwtC6OG2xF7mCv7P7E= -go.opentelemetry.io/proto/otlp v0.7.0/go.mod h1:PqfVotwruBrMGOCsRd/89rSnXhoiJIqeYNgFYFoEGnI= go.uber.org/atomic v1.7.0/go.mod h1:fEN4uk6kAWBTFdckzkM89CLk9XfWZrxpCo0nPH17wJc= go.uber.org/atomic v1.9.0 h1:ECmE8Bn/WFTYwEW/bpKD3M8VtR/zQVbavAoalC1PYyE= -go.uber.org/atomic v1.9.0/go.mod h1:fEN4uk6kAWBTFdckzkM89CLk9XfWZrxpCo0nPH17wJc= -go.uber.org/goleak v1.1.10/go.mod h1:8a7PlsEVH3e/a/GLqe5IIrQx6GzcnRmZEufDUTk4A7A= go.uber.org/multierr v1.6.0/go.mod h1:cdWPpRnG4AhwMwsgIHip0KRBQjJy5kYEpYjJxpXp9iU= go.uber.org/multierr v1.7.0 h1:zaiO/rmgFjbmCXdSYJWQcdvOCsthmdaHfr3Gm2Kx4Ec= -go.uber.org/multierr v1.7.0/go.mod h1:7EAYxJLBy9rStEaz58O2t4Uvip6FSURkq8/ppBp95ak= go.uber.org/zap v1.17.0/go.mod h1:MXVU+bhUf/A7Xi2HNOnopQOrmycQ5Ih87HtOu4q5SSo= go.uber.org/zap v1.19.0 h1:mZQZefskPPCMIBCSEH0v2/iUqqLrYtaeqwD6FUGUnFE= -go.uber.org/zap v1.19.0/go.mod h1:xg/QME4nWcxGxrpdeYfq7UvYrLh66cuVKdrbD1XF/NI= golang.org/x/crypto v0.0.0-20180904163835-0709b304e793/go.mod h1:6SG95UA2DQfeDnfUPMdvaQW0Q7yPrPDi9nlGo2tz2b4= -golang.org/x/crypto v0.0.0-20190228161510-8dd112bcdc25/go.mod h1:djNgcEr1/C05ACkg1iLfiJU5Ep61QUkGW8qpdssI0+w= golang.org/x/crypto v0.0.0-20190308221718-c2843e01d9a2/go.mod h1:djNgcEr1/C05ACkg1iLfiJU5Ep61QUkGW8qpdssI0+w= golang.org/x/crypto v0.0.0-20190510104115-cbcb75029529/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= golang.org/x/crypto v0.0.0-20190605123033-f99c8df09eb5/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= golang.org/x/crypto v0.0.0-20190820162420-60c769a6c586/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= golang.org/x/crypto v0.0.0-20191011191535-87dc89f01550/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI= golang.org/x/crypto v0.0.0-20200622213623-75b288015ac9/go.mod h1:LzIPMQfyMNhhGPhUkYOs5KpL4U8rLKemX1yGLhDgUto= -golang.org/x/crypto v0.0.0-20201002170205-7f63de1d35b0/go.mod h1:LzIPMQfyMNhhGPhUkYOs5KpL4U8rLKemX1yGLhDgUto= +golang.org/x/crypto v0.7.0 h1:AvwMYaRytfdeVt3u6mLaxYtErKYjxA2OXjJ1HHq6t3A= golang.org/x/exp v0.0.0-20190121172915-509febef88a4/go.mod h1:CJ0aWSM057203Lf6IL+f9T1iT9GByDxfZKAQTCR3kQA= golang.org/x/exp v0.0.0-20190306152737-a1d7652674e8/go.mod h1:CJ0aWSM057203Lf6IL+f9T1iT9GByDxfZKAQTCR3kQA= golang.org/x/exp v0.0.0-20190510132918-efd6b22b2522/go.mod h1:ZjyILWgesfNpC6sMxTJOJm9Kp84zZh5NQWvqDGG3Qr8= @@ -461,7 +350,6 @@ golang.org/x/lint v0.0.0-20190930215403-16217165b5de/go.mod h1:6SW0HCj/g11FgYtHl golang.org/x/lint v0.0.0-20191125180803-fdd1cda4f05f/go.mod h1:5qLYkcX4OjUUV8bRuDixDT3tpyyb+LUpUlRWLxfhWrs= golang.org/x/lint v0.0.0-20200130185559-910be7a94367/go.mod h1:3xt1FjdF8hUf6vQPIChWIBhFzV8gjjsPE/fR3IyQdNY= golang.org/x/lint v0.0.0-20200302205851-738671d3881b/go.mod h1:3xt1FjdF8hUf6vQPIChWIBhFzV8gjjsPE/fR3IyQdNY= -golang.org/x/lint v0.0.0-20201208152925-83fdc39ff7b5/go.mod h1:3xt1FjdF8hUf6vQPIChWIBhFzV8gjjsPE/fR3IyQdNY= golang.org/x/lint v0.0.0-20210508222113-6edffad5e616/go.mod h1:3xt1FjdF8hUf6vQPIChWIBhFzV8gjjsPE/fR3IyQdNY= golang.org/x/mobile v0.0.0-20190312151609-d3739f865fa6/go.mod h1:z+o9i4GpDbdi3rU15maQ/Ox0txvL9dWGYEHz965HBQE= golang.org/x/mobile v0.0.0-20190719004257-d2bd2a29d028/go.mod h1:E/iHnbuqvinMTCcRqshq8CkpyQDoeVncDDYHnLhea+o= @@ -471,18 +359,14 @@ golang.org/x/mod v0.1.1-0.20191105210325-c90efee705ee/go.mod h1:QqPTAvyqsEbceGzB golang.org/x/mod v0.1.1-0.20191107180719-034126e5016b/go.mod h1:QqPTAvyqsEbceGzBzNggFXnrqF1CaUcvgkdR5Ot7KZg= golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.4.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= -golang.org/x/mod v0.4.1/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/mod v0.4.2/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA= golang.org/x/net v0.0.0-20180724234803-3673e40ba225/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20180826012351-8a410e7b638d/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= -golang.org/x/net v0.0.0-20180906233101-161cd47e91fd/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20181114220301-adae6a3d119a/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20181220203305-927f97764cc3/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20190108225652-1e06a53dbb7e/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20190213061140-3a22650c66bd/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20190311183353-d8887717615a/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= -golang.org/x/net v0.0.0-20190328230028-74de082e2cca/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190501004415-9ce7a6920f09/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190503192946-f4e77d36d62c/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= @@ -491,7 +375,6 @@ golang.org/x/net v0.0.0-20190613194153-d28f0bde5980/go.mod h1:z5CRVTTTmAJ677TzLL golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20190628185345-da137c7871d7/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20190724013045-ca1201d0de80/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= -golang.org/x/net v0.0.0-20191112182307-2180aed22343/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20191209160850-c0dbc17a3553/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200114155413-6afb5195e5aa/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200202094626-16171245cfb2/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= @@ -507,38 +390,18 @@ golang.org/x/net v0.0.0-20200625001655-4c5254603344/go.mod h1:/O7V0waA8r7cgGh81R golang.org/x/net v0.0.0-20200707034311-ab3426394381/go.mod h1:/O7V0waA8r7cgGh81Ro3o1hOxt32SMVPicZroKQ2sZA= golang.org/x/net v0.0.0-20200822124328-c89045814202/go.mod h1:/O7V0waA8r7cgGh81Ro3o1hOxt32SMVPicZroKQ2sZA= golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= -golang.org/x/net v0.0.0-20201031054903-ff519b6c9102/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= -golang.org/x/net v0.0.0-20201110031124-69a78807bb2b/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= -golang.org/x/net v0.0.0-20201202161906-c7110b5ffcbb/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= -golang.org/x/net v0.0.0-20201209123823-ac852fbbde11/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg= -golang.org/x/net v0.0.0-20210119194325-5f4716e94777/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg= -golang.org/x/net v0.0.0-20210226172049-e18ecbb05110/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg= -golang.org/x/net v0.0.0-20210316092652-d523dce5a7f4/go.mod h1:RBQZq4jEuRlivfhVLdyRGr576XBO4/greRjx4P4O3yc= golang.org/x/net v0.0.0-20210405180319-a5a99cb37ef4/go.mod h1:p54w0d4576C0XHj96bSt6lcn1PtDYWL6XObtHCRCNQM= -golang.org/x/net v0.0.0-20210503060351-7fd8e65b6420/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y= golang.org/x/net v0.0.0-20210525063256-abc453219eb5/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y= -golang.org/x/net v0.0.0-20210614182718-04defd469f4e/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y= -golang.org/x/net v0.0.0-20210825183410-e898025ed96a/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y= golang.org/x/net v0.0.0-20220127200216-cd36cc0744dd/go.mod h1:CfG3xpIq0wQ8r1q4Su4UZFWDARRcnwPjda9FqA0JpMk= golang.org/x/net v0.0.0-20220225172249-27dd8689420f/go.mod h1:CfG3xpIq0wQ8r1q4Su4UZFWDARRcnwPjda9FqA0JpMk= golang.org/x/net v0.8.0 h1:Zrh2ngAOFYneWTAIAPethzeaQLuHwhuBkuV6ZiRnUaQ= golang.org/x/net v0.8.0/go.mod h1:QVkue5JL9kW//ek3r6jTKnTFis1tRmNAW2P1shuFdJc= golang.org/x/oauth2 v0.0.0-20180821212333-d2e6202438be/go.mod h1:N/0e6XlmueqKjAGxoOufVs8QHGRruUQn6yWY3a++T0U= golang.org/x/oauth2 v0.0.0-20190226205417-e64efc72b421/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw= -golang.org/x/oauth2 v0.0.0-20190402181905-9f3314589c9a/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw= golang.org/x/oauth2 v0.0.0-20190604053449-0f29369cfe45/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw= golang.org/x/oauth2 v0.0.0-20191202225959-858c2ad4c8b6/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw= golang.org/x/oauth2 v0.0.0-20200107190931-bf48bf16ab8d/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw= -golang.org/x/oauth2 v0.0.0-20200902213428-5d25da1a8d43/go.mod h1:KelEdhl1UZF7XfJ4dDtk6s++YSgaE7mD/BuKKDLBl4A= -golang.org/x/oauth2 v0.0.0-20201109201403-9fd604954f58/go.mod h1:KelEdhl1UZF7XfJ4dDtk6s++YSgaE7mD/BuKKDLBl4A= -golang.org/x/oauth2 v0.0.0-20201208152858-08078c50e5b5/go.mod h1:KelEdhl1UZF7XfJ4dDtk6s++YSgaE7mD/BuKKDLBl4A= -golang.org/x/oauth2 v0.0.0-20210218202405-ba52d332ba99/go.mod h1:KelEdhl1UZF7XfJ4dDtk6s++YSgaE7mD/BuKKDLBl4A= -golang.org/x/oauth2 v0.0.0-20210220000619-9bb904979d93/go.mod h1:KelEdhl1UZF7XfJ4dDtk6s++YSgaE7mD/BuKKDLBl4A= -golang.org/x/oauth2 v0.0.0-20210313182246-cd4f82c27b84/go.mod h1:KelEdhl1UZF7XfJ4dDtk6s++YSgaE7mD/BuKKDLBl4A= golang.org/x/oauth2 v0.0.0-20210514164344-f6687ab2804c/go.mod h1:KelEdhl1UZF7XfJ4dDtk6s++YSgaE7mD/BuKKDLBl4A= -golang.org/x/oauth2 v0.0.0-20210628180205-a41e5a781914/go.mod h1:KelEdhl1UZF7XfJ4dDtk6s++YSgaE7mD/BuKKDLBl4A= -golang.org/x/oauth2 v0.0.0-20210805134026-6f1e6394065a/go.mod h1:KelEdhl1UZF7XfJ4dDtk6s++YSgaE7mD/BuKKDLBl4A= -golang.org/x/oauth2 v0.0.0-20210819190943-2bc19b11175f/go.mod h1:KelEdhl1UZF7XfJ4dDtk6s++YSgaE7mD/BuKKDLBl4A= golang.org/x/oauth2 v0.0.0-20220223155221-ee480838109b/go.mod h1:DAh4E804XQdzx2j+YRIaUnCqCV2RuMz24cGBJ5QYIrc= golang.org/x/oauth2 v0.4.0 h1:NF0gk8LVPg1Ml7SSbGyySuoxdsXitj7TvgvuRxIMc/M= golang.org/x/sync v0.0.0-20180314180146-1d60e4601c6f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= @@ -556,7 +419,6 @@ golang.org/x/sync v0.0.0-20220601150217-0de741cfad7f h1:Ax0t5p6N38Ga0dThY21weqDE golang.org/x/sync v0.0.0-20220601150217-0de741cfad7f/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM= golang.org/x/sys v0.0.0-20180830151530-49385e6e1522/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= golang.org/x/sys v0.0.0-20180905080454-ebe1bf3edb33/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= -golang.org/x/sys v0.0.0-20180909124046-d0be0721c37e/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= golang.org/x/sys v0.0.0-20181107165924-66b7b1311ac8/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= golang.org/x/sys v0.0.0-20181116152217-5ac8a444bdc5/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= golang.org/x/sys v0.0.0-20190215142949-d0b11bdaac8a/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= @@ -569,8 +431,6 @@ golang.org/x/sys v0.0.0-20190606165138-5da285871e9c/go.mod h1:h1NjWce9XRLGQEsW7w golang.org/x/sys v0.0.0-20190624142023-c5567b49c5d0/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20190726091711-fc99dfbffb4e/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20191001151750-bb3f8db39f24/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -golang.org/x/sys v0.0.0-20191026070338-33540a1f6037/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -golang.org/x/sys v0.0.0-20191112214154-59a1497f0cea/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20191204072324-ce4227a45e2e/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20191228213918-04cbcbbfeed8/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20200106162015-b016eb3dc98e/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= @@ -589,31 +449,15 @@ golang.org/x/sys v0.0.0-20200523222454-059865788121/go.mod h1:h1NjWce9XRLGQEsW7w golang.org/x/sys v0.0.0-20200615200032-f1bc736245b1/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20200625212154-ddb9806d33ae/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20200803210538-64077c9b5642/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -golang.org/x/sys v0.0.0-20200828194041-157a740278f4/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -golang.org/x/sys v0.0.0-20200905004654-be1d3432aa8f/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20200930185726-fdedc70b468f/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20201119102817-f84b799fce68/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -golang.org/x/sys v0.0.0-20201201145000-ef89a241ccb3/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -golang.org/x/sys v0.0.0-20210104204734-6f8348627aad/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -golang.org/x/sys v0.0.0-20210119212857-b64e53b001e4/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20210124154548-22da62e12c0c/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -golang.org/x/sys v0.0.0-20210220050731-9a76102bfb43/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -golang.org/x/sys v0.0.0-20210305230114-8fe3ee5dd75b/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -golang.org/x/sys v0.0.0-20210315160823-c6e025ad8005/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= -golang.org/x/sys v0.0.0-20210320140829-1e4c9ba3b0c4/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20210330210617-4fbd30eecc44/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20210403161142-5e06dd20ab57/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20210423082822-04245dca01da/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= golang.org/x/sys v0.0.0-20210510120138-977fb7262007/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= -golang.org/x/sys v0.0.0-20210514084401-e8d321eab015/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= golang.org/x/sys v0.0.0-20210603081109-ebe580a85c40/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= -golang.org/x/sys v0.0.0-20210603125802-9665404d3644/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= golang.org/x/sys v0.0.0-20210615035016-665e8c7367d1/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= -golang.org/x/sys v0.0.0-20210616094352-59db8d763f22/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= -golang.org/x/sys v0.0.0-20210630005230-0f9fa26af87c/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= -golang.org/x/sys v0.0.0-20210806184541-e5e7981a1069/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= -golang.org/x/sys v0.0.0-20210823070655-63515b42dcdf/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= -golang.org/x/sys v0.0.0-20210831042530-f4d43177bf5e/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= golang.org/x/sys v0.0.0-20211216021012-1d35b9e2eb4e/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= golang.org/x/sys v0.0.0-20220114195835-da31bd327af9/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= golang.org/x/sys v0.0.0-20220715151400-c0bba94af5f8/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= @@ -624,22 +468,16 @@ golang.org/x/term v0.0.0-20210927222741-03fcf44c2211/go.mod h1:jbD1KX2456YbFQfuX golang.org/x/text v0.0.0-20170915032832-14c0d48ead0c/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ= golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ= golang.org/x/text v0.3.1-0.20180807135948-17ff2d5776d2/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ= -golang.org/x/text v0.3.1-0.20181227161524-e6919f6577db/go.mod h1:bEr9sfX3Q8Zfm5fL9x+3itogRgK3+ptLWKqgva+5dAk= golang.org/x/text v0.3.2/go.mod h1:bEr9sfX3Q8Zfm5fL9x+3itogRgK3+ptLWKqgva+5dAk= golang.org/x/text v0.3.3/go.mod h1:5Zoc/QRtKVWzQhOtBMvqHzDpF6irO9z98xDceosuGiQ= -golang.org/x/text v0.3.4/go.mod h1:5Zoc/QRtKVWzQhOtBMvqHzDpF6irO9z98xDceosuGiQ= golang.org/x/text v0.3.5/go.mod h1:5Zoc/QRtKVWzQhOtBMvqHzDpF6irO9z98xDceosuGiQ= golang.org/x/text v0.3.6/go.mod h1:5Zoc/QRtKVWzQhOtBMvqHzDpF6irO9z98xDceosuGiQ= golang.org/x/text v0.3.7/go.mod h1:u+2+/6zg+i71rQMx5EYifcz6MCKuco9NR6JIITiCfzQ= golang.org/x/text v0.8.0 h1:57P1ETyNKtuIjB4SRd15iJxuhj8Gc416Y78H3qgMh68= golang.org/x/text v0.8.0/go.mod h1:e1OnstbJyHTd6l/uOt8jFFHp6TRDWZR/bV3emEE/zU8= -golang.org/x/time v0.0.0-20161028155119-f51c12702a4d/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ= -golang.org/x/time v0.0.0-20180412165947-fbb02b2291d2/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ= golang.org/x/time v0.0.0-20181108054448-85acf8d2951c/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ= golang.org/x/time v0.0.0-20190308202827-9d24e82272b4/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ= golang.org/x/time v0.0.0-20191024005414-555d28b269f0/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ= -golang.org/x/time v0.0.0-20200630173020-3af7569d3a1e/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ= -golang.org/x/tools v0.0.0-20180221164845-07fd8470d635/go.mod h1:n7NCudcB/nEzxVGmLbDWY5pfWTLqBcC2KZ6jyYvM4mQ= golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGmLbDWY5pfWTLqBcC2KZ6jyYvM4mQ= golang.org/x/tools v0.0.0-20190114222345-bf090417da8b/go.mod h1:n7NCudcB/nEzxVGmLbDWY5pfWTLqBcC2KZ6jyYvM4mQ= golang.org/x/tools v0.0.0-20190226205152-f727befe758c/go.mod h1:9Yl7xja0Znq3iFh3HoIrodX9oNMXvdceNzlUR8zjMvY= @@ -655,7 +493,6 @@ golang.org/x/tools v0.0.0-20190628153133-6cdbf07be9d0/go.mod h1:/rFqwRUd4F7ZHNgw golang.org/x/tools v0.0.0-20190816200558-6889da9d5479/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20190911174233-4f2ddba30aff/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20191012152004-8de300cfc20a/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= -golang.org/x/tools v0.0.0-20191108193012-7d206e10da11/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20191113191852-77e3bb0ad9e7/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20191115202509-3a792d9c32b2/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= golang.org/x/tools v0.0.0-20191119224855-298f0cb1881e/go.mod h1:b+2E5dAYhXwXZwtnZ6UAqBI28+e2cm9otk0dWdXHAEo= @@ -682,18 +519,8 @@ golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roY golang.org/x/tools v0.0.0-20200729194436-6467de6f59a7/go.mod h1:njjCfa9FT2d7l9Bc6FUM5FLjQPp3cFF28FI3qnDFljA= golang.org/x/tools v0.0.0-20200804011535-6c149bb5ef0d/go.mod h1:njjCfa9FT2d7l9Bc6FUM5FLjQPp3cFF28FI3qnDFljA= golang.org/x/tools v0.0.0-20200825202427-b303f430e36d/go.mod h1:njjCfa9FT2d7l9Bc6FUM5FLjQPp3cFF28FI3qnDFljA= -golang.org/x/tools v0.0.0-20200904185747-39188db58858/go.mod h1:Cj7w3i3Rnn0Xh82ur9kSqwfTHTeVxaDqrfMjpcNT6bE= -golang.org/x/tools v0.0.0-20201110124207-079ba7bd75cd/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.0.0-20201201161351-ac6f37ff4c2a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.0.0-20201208233053-a543418bbed2/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.0.0-20210105154028-b0ab187a4818/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA= -golang.org/x/tools v0.1.0/go.mod h1:xkSsbof2nBLbhDlRMhhhyNLN/zl3eTqcnHD5viDpcZ0= -golang.org/x/tools v0.1.1/go.mod h1:o0xws9oXOQQZyjljx8fwUC0k7L1pTE6eaCbjGeHmOkk= golang.org/x/tools v0.1.2/go.mod h1:o0xws9oXOQQZyjljx8fwUC0k7L1pTE6eaCbjGeHmOkk= -golang.org/x/tools v0.1.3/go.mod h1:o0xws9oXOQQZyjljx8fwUC0k7L1pTE6eaCbjGeHmOkk= -golang.org/x/tools v0.1.4/go.mod h1:o0xws9oXOQQZyjljx8fwUC0k7L1pTE6eaCbjGeHmOkk= -golang.org/x/tools v0.1.5/go.mod h1:o0xws9oXOQQZyjljx8fwUC0k7L1pTE6eaCbjGeHmOkk= golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= @@ -715,18 +542,6 @@ google.golang.org/api v0.24.0/go.mod h1:lIXQywCXRcnZPGlsd8NbLnOjtAoL6em04bJ9+z0M google.golang.org/api v0.28.0/go.mod h1:lIXQywCXRcnZPGlsd8NbLnOjtAoL6em04bJ9+z0MncE= google.golang.org/api v0.29.0/go.mod h1:Lcubydp8VUV7KeIHD9z2Bys/sm/vGKnG1UHuDBSrHWM= google.golang.org/api v0.30.0/go.mod h1:QGmEvQ87FHZNiUVJkT14jQNYJ4ZJjdRF23ZXz5138Fc= -google.golang.org/api v0.35.0/go.mod h1:/XrVsuzM0rZmrsbjJutiuftIzeuTQcEeaYcSk/mQ1dg= -google.golang.org/api v0.36.0/go.mod h1:+z5ficQTmoYpPn8LCUNVpK5I7hwkpjbcgqA7I34qYtE= -google.golang.org/api v0.40.0/go.mod h1:fYKFpnQN0DsDSKRVRcQSDQNtqWPfM9i+zNPxepjRCQ8= -google.golang.org/api v0.41.0/go.mod h1:RkxM5lITDfTzmyKFPt+wGrCJbVfniCr2ool8kTBzRTU= -google.golang.org/api v0.43.0/go.mod h1:nQsDGjRXMo4lvh5hP0TKqF244gqhGcr/YSIykhUk/94= -google.golang.org/api v0.47.0/go.mod h1:Wbvgpq1HddcWVtzsVLyfLp8lDg6AA241LmgIL59tHXo= -google.golang.org/api v0.48.0/go.mod h1:71Pr1vy+TAZRPkPs/xlCf5SsU8WjuAWv1Pfjbtukyy4= -google.golang.org/api v0.50.0/go.mod h1:4bNT5pAuq5ji4SRZm+5QIkjny9JAyVD/3gaSihNefaw= -google.golang.org/api v0.51.0/go.mod h1:t4HdrdoNgyN5cbEfm7Lum0lcLDLiise1F8qDKX00sOU= -google.golang.org/api v0.54.0/go.mod h1:7C4bFFOvVDGXjfDTAsgGwDgAxRDeQ4X8NvUedIt6z3k= -google.golang.org/api v0.55.0/go.mod h1:38yMfeP1kfjsl8isn0tliTjIb1rJXcQi4UXlbqivdVE= -google.golang.org/api v0.56.0/go.mod h1:38yMfeP1kfjsl8isn0tliTjIb1rJXcQi4UXlbqivdVE= google.golang.org/api v0.60.0 h1:eq/zs5WPH4J9undYM9IP1O7dSr7Yh8Y0GtSCpzGzIUk= google.golang.org/appengine v1.1.0/go.mod h1:EbEs0AVv82hx2wNQdGPgUI5lhzA/G0D9YwlJXL52JkM= google.golang.org/appengine v1.4.0/go.mod h1:xpcJRLb0r/rnEns0DIKYYv+WjYCduHsrkT7/EB5XEv4= @@ -735,7 +550,6 @@ google.golang.org/appengine v1.6.1/go.mod h1:i06prIuMbXzDqacNJfV5OdTW448YApPu5ww google.golang.org/appengine v1.6.5/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc= google.golang.org/appengine v1.6.6/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc= google.golang.org/appengine v1.6.7 h1:FZR1q0exgwxzPzp/aF+VccGrSfxfPpkBqjIIEq3ru6c= -google.golang.org/appengine v1.6.7/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc= google.golang.org/genproto v0.0.0-20180817151627-c66870c02cf8/go.mod h1:JiN7NxoALGmiZfu7CAH4rXhgtRTLTxftemlI0sWmxmc= google.golang.org/genproto v0.0.0-20190307195333-5fe7a883aa19/go.mod h1:VzzqZJRnGkLBvHegQrXjBqPurQTc5/KpmUdxsrq26oE= google.golang.org/genproto v0.0.0-20190418145605-e7d98fc518a7/go.mod h1:VzzqZJRnGkLBvHegQrXjBqPurQTc5/KpmUdxsrq26oE= @@ -766,37 +580,13 @@ google.golang.org/genproto v0.0.0-20200618031413-b414f8b61790/go.mod h1:jDfRM7Fc google.golang.org/genproto v0.0.0-20200729003335-053ba62fc06f/go.mod h1:FWY/as6DDZQgahTzZj3fqbO1CbirC29ZNUFHwi0/+no= google.golang.org/genproto v0.0.0-20200804131852-c06518451d9c/go.mod h1:FWY/as6DDZQgahTzZj3fqbO1CbirC29ZNUFHwi0/+no= google.golang.org/genproto v0.0.0-20200825200019-8632dd797987/go.mod h1:FWY/as6DDZQgahTzZj3fqbO1CbirC29ZNUFHwi0/+no= -google.golang.org/genproto v0.0.0-20200904004341-0bd0a958aa1d/go.mod h1:FWY/as6DDZQgahTzZj3fqbO1CbirC29ZNUFHwi0/+no= -google.golang.org/genproto v0.0.0-20201109203340-2640f1f9cdfb/go.mod h1:FWY/as6DDZQgahTzZj3fqbO1CbirC29ZNUFHwi0/+no= -google.golang.org/genproto v0.0.0-20201201144952-b05cb90ed32e/go.mod h1:FWY/as6DDZQgahTzZj3fqbO1CbirC29ZNUFHwi0/+no= -google.golang.org/genproto v0.0.0-20201210142538-e3217bee35cc/go.mod h1:FWY/as6DDZQgahTzZj3fqbO1CbirC29ZNUFHwi0/+no= -google.golang.org/genproto v0.0.0-20201214200347-8c77b98c765d/go.mod h1:FWY/as6DDZQgahTzZj3fqbO1CbirC29ZNUFHwi0/+no= -google.golang.org/genproto v0.0.0-20210222152913-aa3ee6e6a81c/go.mod h1:FWY/as6DDZQgahTzZj3fqbO1CbirC29ZNUFHwi0/+no= -google.golang.org/genproto v0.0.0-20210303154014-9728d6b83eeb/go.mod h1:FWY/as6DDZQgahTzZj3fqbO1CbirC29ZNUFHwi0/+no= -google.golang.org/genproto v0.0.0-20210310155132-4ce2db91004e/go.mod h1:FWY/as6DDZQgahTzZj3fqbO1CbirC29ZNUFHwi0/+no= -google.golang.org/genproto v0.0.0-20210319143718-93e7006c17a6/go.mod h1:FWY/as6DDZQgahTzZj3fqbO1CbirC29ZNUFHwi0/+no= -google.golang.org/genproto v0.0.0-20210402141018-6c239bbf2bb1/go.mod h1:9lPAdzaEmUacj36I+k7YKbEc5CXzPIeORRgDAUOu28A= -google.golang.org/genproto v0.0.0-20210513213006-bf773b8c8384/go.mod h1:P3QM42oQyzQSnHPnZ/vqoCdDmzH28fzWByN9asMeM8A= google.golang.org/genproto v0.0.0-20210602131652-f16073e35f0c/go.mod h1:UODoCrxHCcBojKKwX1terBiRUaqAsFqJiF615XL43r0= -google.golang.org/genproto v0.0.0-20210604141403-392c879c8b08/go.mod h1:UODoCrxHCcBojKKwX1terBiRUaqAsFqJiF615XL43r0= -google.golang.org/genproto v0.0.0-20210608205507-b6d2f5bf0d7d/go.mod h1:UODoCrxHCcBojKKwX1terBiRUaqAsFqJiF615XL43r0= -google.golang.org/genproto v0.0.0-20210624195500-8bfb893ecb84/go.mod h1:SzzZ/N+nwJDaO1kznhnlzqS8ocJICar6hYhVyhi++24= -google.golang.org/genproto v0.0.0-20210713002101-d411969a0d9a/go.mod h1:AxrInvYm1dci+enl5hChSFPOmmUF1+uAa/UsgNRWd7k= -google.golang.org/genproto v0.0.0-20210716133855-ce7ef5c701ea/go.mod h1:AxrInvYm1dci+enl5hChSFPOmmUF1+uAa/UsgNRWd7k= -google.golang.org/genproto v0.0.0-20210728212813-7823e685a01f/go.mod h1:ob2IJxKrgPT52GcgX759i1sleT07tiKowYBGbczaW48= -google.golang.org/genproto v0.0.0-20210805201207-89edb61ffb67/go.mod h1:ob2IJxKrgPT52GcgX759i1sleT07tiKowYBGbczaW48= -google.golang.org/genproto v0.0.0-20210813162853-db860fec028c/go.mod h1:cFeNkxwySK631ADgubI+/XFU/xp8FD5KIVV4rj8UC5w= -google.golang.org/genproto v0.0.0-20210821163610-241b8fcbd6c8/go.mod h1:eFjDcFEctNawg4eG61bRv87N7iHBWyVhJu7u1kqDUXY= -google.golang.org/genproto v0.0.0-20210825212027-de86158e7fda/go.mod h1:eFjDcFEctNawg4eG61bRv87N7iHBWyVhJu7u1kqDUXY= -google.golang.org/genproto v0.0.0-20210828152312-66f60bf46e71/go.mod h1:eFjDcFEctNawg4eG61bRv87N7iHBWyVhJu7u1kqDUXY= -google.golang.org/genproto v0.0.0-20210831024726-fe130286e0e2/go.mod h1:eFjDcFEctNawg4eG61bRv87N7iHBWyVhJu7u1kqDUXY= google.golang.org/genproto v0.0.0-20230110181048-76db0878b65f h1:BWUVssLB0HVOSY78gIdvk1dTVYtT1y8SBWtPYuTJ/6w= google.golang.org/genproto v0.0.0-20230110181048-76db0878b65f/go.mod h1:RGgjbofJ8xD9Sq1VVhDM1Vok1vRONV+rg+CjzG4SZKM= google.golang.org/grpc v1.19.0/go.mod h1:mqu4LbDTu4XGKhr4mRzUsmM4RtVoemTSY81AxZiDr8c= google.golang.org/grpc v1.19.1/go.mod h1:mqu4LbDTu4XGKhr4mRzUsmM4RtVoemTSY81AxZiDr8c= google.golang.org/grpc v1.20.1/go.mod h1:10oTOabMzJvdu6/UiuZezV6QK5dSlG84ov/aaiqXj38= google.golang.org/grpc v1.21.1/go.mod h1:oYelfM1adQP15Ek0mdvEgi9Df8B9CZIaU1084ijfRaM= -google.golang.org/grpc v1.22.0/go.mod h1:Y5yQAOtifL1yxbo5wqy6BxZv8vAUGQwXBOALyacEbxg= google.golang.org/grpc v1.23.0/go.mod h1:Y5yQAOtifL1yxbo5wqy6BxZv8vAUGQwXBOALyacEbxg= google.golang.org/grpc v1.25.1/go.mod h1:c3i+UQWmh7LiEpx4sFZnkU36qjEYZ0imhYfXVyQciAY= google.golang.org/grpc v1.26.0/go.mod h1:qbnxyOmOxrQa7FizSgH+ReBfzJrCY1pSN7KXBS8abTk= @@ -806,22 +596,10 @@ google.golang.org/grpc v1.28.0/go.mod h1:rpkK4SK4GF4Ach/+MFLZUBavHOvF2JJB5uozKKa google.golang.org/grpc v1.29.1/go.mod h1:itym6AZVZYACWQqET3MqgPpjcuV5QH3BxFS3IjizoKk= google.golang.org/grpc v1.30.0/go.mod h1:N36X2cJ7JwdamYAgDz+s+rVMFjt3numwzf/HckM8pak= google.golang.org/grpc v1.31.0/go.mod h1:N36X2cJ7JwdamYAgDz+s+rVMFjt3numwzf/HckM8pak= -google.golang.org/grpc v1.31.1/go.mod h1:N36X2cJ7JwdamYAgDz+s+rVMFjt3numwzf/HckM8pak= google.golang.org/grpc v1.33.1/go.mod h1:fr5YgcSWrqhRRxogOsw7RzIpsmvOZ6IcH4kBYTpR3n0= -google.golang.org/grpc v1.33.2/go.mod h1:JMHMWHQWaTccqQQlmk3MJZS+GWXOdAesneDmEnv2fbc= -google.golang.org/grpc v1.34.0/go.mod h1:WotjhfgOW/POjDeRt8vscBtXq+2VjORFy659qA51WJ8= -google.golang.org/grpc v1.35.0/go.mod h1:qjiiYl8FncCW8feJPdyg3v6XW24KsRHe+dy9BAGRRjU= -google.golang.org/grpc v1.36.0/go.mod h1:qjiiYl8FncCW8feJPdyg3v6XW24KsRHe+dy9BAGRRjU= -google.golang.org/grpc v1.36.1/go.mod h1:qjiiYl8FncCW8feJPdyg3v6XW24KsRHe+dy9BAGRRjU= -google.golang.org/grpc v1.37.0/go.mod h1:NREThFqKR1f3iQ6oBuvc5LadQuXVGo9rkm5ZGrQdJfM= -google.golang.org/grpc v1.37.1/go.mod h1:NREThFqKR1f3iQ6oBuvc5LadQuXVGo9rkm5ZGrQdJfM= google.golang.org/grpc v1.38.0/go.mod h1:NREThFqKR1f3iQ6oBuvc5LadQuXVGo9rkm5ZGrQdJfM= -google.golang.org/grpc v1.39.0/go.mod h1:PImNr+rS9TWYb2O4/emRugxiyHZ5JyHW5F+RPnDzfrE= -google.golang.org/grpc v1.39.1/go.mod h1:PImNr+rS9TWYb2O4/emRugxiyHZ5JyHW5F+RPnDzfrE= -google.golang.org/grpc v1.40.0/go.mod h1:ogyxbiOoUXAkP+4+xa6PZSE9DZgIHtSpzjDTB9KAK34= google.golang.org/grpc v1.53.0 h1:LAv2ds7cmFV/XTS3XG1NneeENYrXGmorPxsBbptIjNc= google.golang.org/grpc v1.53.0/go.mod h1:OnIrk0ipVdj4N5d9IUoFUx72/VlD7+jUsHwZgwSMQpw= -google.golang.org/grpc/cmd/protoc-gen-go-grpc v1.1.0/go.mod h1:6Kw0yEErY5E/yWrBtf03jp27GLLJujG4z/JK95pnjjw= google.golang.org/protobuf v0.0.0-20200109180630-ec00e32a8dfd/go.mod h1:DFci5gLYBciE7Vtevhsrf46CRTquxDuWsQurQQe4oz8= google.golang.org/protobuf v0.0.0-20200221191635-4d8936d0db64/go.mod h1:kwYJMbMJ01Woi6D6+Kah6886xMZcty6N08ah7+eCXa0= google.golang.org/protobuf v0.0.0-20200228230310-ab0ca4ff8a60/go.mod h1:cfTl7dwQJ+fmap5saPgwCLgHXTUD7jkjRqWcaiX5VyM= @@ -834,22 +612,15 @@ google.golang.org/protobuf v1.24.0/go.mod h1:r/3tXBNzIEhYS9I1OUVjXDlt8tc493IdKGj google.golang.org/protobuf v1.25.0/go.mod h1:9JNX74DMeImyA3h4bdi1ymwjUzf21/xIlbajtzgsN7c= google.golang.org/protobuf v1.26.0-rc.1/go.mod h1:jlhhOSvTdKEhbULTjvd4ARK9grFBp09yW+WbY/TyQbw= google.golang.org/protobuf v1.26.0/go.mod h1:9q0QmTI4eRPtz6boOQmLYwt+qCgq0jsYwAQnmE0givc= -google.golang.org/protobuf v1.27.1/go.mod h1:9q0QmTI4eRPtz6boOQmLYwt+qCgq0jsYwAQnmE0givc= google.golang.org/protobuf v1.28.1 h1:d0NfwRgPtno5B1Wa6L2DAG+KivqkdutMf1UhdNx175w= google.golang.org/protobuf v1.28.1/go.mod h1:HV8QOd/L58Z+nl8r43ehVNZIU/HEI6OcFqwMG9pJV4I= -gopkg.in/airbrake/gobrake.v2 v2.0.9/go.mod h1:/h5ZAUhDkGaJfjzjKLSjv6zCL6O0LLBxU4K+aSYdM/U= gopkg.in/alecthomas/kingpin.v2 v2.2.6/go.mod h1:FMv+mEhP44yOT+4EoQTLFTRgOQ1FBLkstjWtayDeSgw= gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= gopkg.in/check.v1 v1.0.0-20180628173108-788fd7840127/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= gopkg.in/check.v1 v1.0.0-20190902080502-41f04d3bba15/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= gopkg.in/check.v1 v1.0.0-20201130134442-10cb98267c6c h1:Hei/4ADfdWqJk1ZMxUNpqntNwaWcugrBjAiHlqqRiVk= -gopkg.in/check.v1 v1.0.0-20201130134442-10cb98267c6c/go.mod h1:JHkPIbrfpd72SG/EVd6muEfDQjcINNoR0C8j2r3qZ4Q= gopkg.in/errgo.v2 v2.1.0/go.mod h1:hNsd1EY+bozCKY1Ytp96fpM3vjJbqLJn88ws8XvfDNI= -gopkg.in/fsnotify.v1 v1.4.7/go.mod h1:Tz8NjZHkW78fSQdbUxIjBTcgA1z1m8ZHf0WmKUhAMys= -gopkg.in/gemnasium/logrus-airbrake-hook.v2 v2.1.2/go.mod h1:Xk6kEKp8OKb+X14hQBKWaSkCsqBpgog8nAV2xsGOxlo= -gopkg.in/inf.v0 v0.9.0/go.mod h1:cWUDdTG/fYaXco+Dcufb5Vnc6Gp2YChqWtbxRZE0mXw= gopkg.in/resty.v1 v1.12.0/go.mod h1:mDo4pnntr5jdWRML875a/NmxYqAlA73dVijT2AXvQQo= -gopkg.in/tomb.v1 v1.0.0-20141024135613-dd632973f1e7/go.mod h1:dt/ZhP58zS4L8KSrWDmTeBkI65Dw0HsyUHuEVlX15mw= gopkg.in/yaml.v2 v2.0.0-20170812160011-eb3733d160e7/go.mod h1:JAlM8MvJe8wmxCU4Bli9HhUf9+ttbYbLASfIpnQbh74= gopkg.in/yaml.v2 v2.2.1/go.mod h1:hI93XBmqTisBFMUTm0b8Fm+jr3Dg1NNxqwp+5A1VGuI= gopkg.in/yaml.v2 v2.2.2/go.mod h1:hI93XBmqTisBFMUTm0b8Fm+jr3Dg1NNxqwp+5A1VGuI= @@ -871,14 +642,7 @@ honnef.co/go/tools v0.0.0-20190523083050-ea95bdfd59fc/go.mod h1:rf3lG4BRIbNafJWh honnef.co/go/tools v0.0.1-2019.2.3/go.mod h1:a3bituU0lyd329TUQxRnasdCoJDkEUEAqEt0JzvZhAg= honnef.co/go/tools v0.0.1-2020.1.3/go.mod h1:X/FiERA/W4tHapMX5mGpAtMSVEeEUOyHaw9vFzvIQ3k= honnef.co/go/tools v0.0.1-2020.1.4/go.mod h1:X/FiERA/W4tHapMX5mGpAtMSVEeEUOyHaw9vFzvIQ3k= -k8s.io/api v0.0.0-20190620073856-dcce3486da33/go.mod h1:ldk709UQo/iedNLOW7J06V9QSSGY5heETKeWqnPoqF8= -k8s.io/apimachinery v0.0.0-20190620073744-d16981aedf33/go.mod h1:9q5NW/mMno/nwbRZd/Ks2TECgi2PTZ9cwarf4q+ze6Q= -k8s.io/client-go v0.0.0-20190620074045-585a16d2e773/go.mod h1:miKCC7C/WGwJqcDctyJtAnP3Gss0Y5KwURqJ7q5pfEw= -k8s.io/klog v0.3.1/go.mod h1:Gq+BEi5rUBO/HRz0bTSXDUcqjScdoY3a9IHpCEIOOfk= -k8s.io/kube-openapi v0.0.0-20190228160746-b3a7cee44a30/go.mod h1:BXM9ceUBTj2QnfH2MK1odQs778ajze1RxcmP6S8RVVc= -k8s.io/utils v0.0.0-20190221042446-c2654d5206da/go.mod h1:8k8uAuAQ0rXslZKaEWd0c3oVhZz7sSzSiPnVZayjIX0= rsc.io/binaryregexp v0.2.0/go.mod h1:qTv7/COck+e2FymRvadv62gMdZztPaShugOCi3I+8D8= rsc.io/quote/v3 v3.1.0/go.mod h1:yEA65RcK8LyAZtP9Kv3t0HmxON59tX3rD+tICJqUlj0= rsc.io/sampler v1.3.0/go.mod h1:T1hPZKmBbMNahiBKFy5HrXp6adAjACjK9JXDnKaTXpA= -sigs.k8s.io/yaml v1.1.0/go.mod h1:UJmg0vDUVViEyp3mgSv9WPwZCDxu4rQW1olrI1uml+o= sigs.k8s.io/yaml v1.2.0/go.mod h1:yfXDCHCao9+ENCvLSE62v9VSji2MKu5jeNfTrofGhJc=