aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorZhongpeng Lin <zplin@uber.com>2023-12-07 07:54:09 -0800
committerGitHub <noreply@github.com>2023-12-07 10:54:09 -0500
commit7e33c12ee96192f12b7fbb089e6fe1af6a29ef14 (patch)
treee72deca781a53f8fb0d901efcc631d9134b2fed3
parent6c32c3b917cc5d3cfee680c03179d7552832bb3f (diff)
downloadbazelbuild-remote-apis-7e33c12ee96192f12b7fbb089e6fe1af6a29ef14.tar.gz
Upgrade rules_go, Gazelle, googleapis, com_google_protobuf (#269)
-rw-r--r--WORKSPACE37
-rw-r--r--build/bazel/remote/asset/v1/BUILD6
-rw-r--r--build/bazel/remote/asset/v1/go/BUILD4
-rw-r--r--build/bazel/remote/execution/v2/BUILD8
-rw-r--r--build/bazel/remote/execution/v2/go/BUILD6
-rwxr-xr-xbuild/bazel/remote/execution/v2/remote_execution.pb.go553
-rw-r--r--external/BUILD.googleapis78
-rw-r--r--remote_apis_deps.bzl18
8 files changed, 333 insertions, 377 deletions
diff --git a/WORKSPACE b/WORKSPACE
index 7ef7349..391d0c2 100644
--- a/WORKSPACE
+++ b/WORKSPACE
@@ -4,10 +4,10 @@ load("@bazel_tools//tools/build_defs/repo:http.bzl", "http_archive")
http_archive(
name = "bazel_skylib",
- sha256 = "97e70364e9249702246c0e9444bccdc4b847bed1eb03c5a3ece4f83dfe6abc44",
+ sha256 = "66ffd9315665bfaafc96b52278f57c7e2dd09f5ede279ea6d39b2be471e7e3aa",
urls = [
- "https://mirror.bazel.build/github.com/bazelbuild/bazel-skylib/releases/download/1.0.2/bazel-skylib-1.0.2.tar.gz",
- "https://github.com/bazelbuild/bazel-skylib/releases/download/1.0.2/bazel-skylib-1.0.2.tar.gz",
+ "https://mirror.bazel.build/github.com/bazelbuild/bazel-skylib/releases/download/1.4.2/bazel-skylib-1.4.2.tar.gz",
+ "https://github.com/bazelbuild/bazel-skylib/releases/download/1.4.2/bazel-skylib-1.4.2.tar.gz",
],
)
@@ -18,29 +18,29 @@ bazel_skylib_workspace()
# Pull in go rules, which we need in order to selectively pull in Go dependencies.
http_archive(
name = "io_bazel_rules_go",
- sha256 = "69de5c704a05ff37862f7e0f5534d4f479418afc21806c887db544a316f3cb6b",
+ sha256 = "d6ab6b57e48c09523e93050f13698f708428cfd5e619252e369d377af6597707",
urls = [
- "https://mirror.bazel.build/github.com/bazelbuild/rules_go/releases/download/v0.27.0/rules_go-v0.27.0.tar.gz",
- "https://github.com/bazelbuild/rules_go/releases/download/v0.27.0/rules_go-v0.27.0.tar.gz",
+ "https://mirror.bazel.build/github.com/bazelbuild/rules_go/releases/download/v0.43.0/rules_go-v0.43.0.zip",
+ "https://github.com/bazelbuild/rules_go/releases/download/v0.43.0/rules_go-v0.43.0.zip",
],
)
# Gazelle, which we need in order to selectively pull in Gazelle dependencies for Go.
http_archive(
name = "bazel_gazelle",
- sha256 = "62ca106be173579c0a167deb23358fdfe71ffa1e4cfdddf5582af26520f1c66f",
+ sha256 = "b7387f72efb59f876e4daae42f1d3912d0d45563eac7cb23d1de0b094ab588cf",
urls = [
- "https://mirror.bazel.build/github.com/bazelbuild/bazel-gazelle/releases/download/v0.23.0/bazel-gazelle-v0.23.0.tar.gz",
- "https://github.com/bazelbuild/bazel-gazelle/releases/download/v0.23.0/bazel-gazelle-v0.23.0.tar.gz",
+ "https://mirror.bazel.build/github.com/bazelbuild/bazel-gazelle/releases/download/v0.34.0/bazel-gazelle-v0.34.0.tar.gz",
+ "https://github.com/bazelbuild/bazel-gazelle/releases/download/v0.34.0/bazel-gazelle-v0.34.0.tar.gz",
],
)
# Needed for protobuf.
http_archive(
name = "com_google_protobuf",
- sha256 = "678d91d8a939a1ef9cb268e1f20c14cd55e40361dc397bb5881e4e1e532679b1",
- strip_prefix = "protobuf-3.10.1",
- urls = ["https://github.com/protocolbuffers/protobuf/archive/v3.10.1.zip"],
+ sha256 = "535fbf566d372ccf3a097c374b26896fa044bf4232aef9cab37bd1cc1ba4e850",
+ strip_prefix = "protobuf-3.15.0",
+ urls = ["https://github.com/protocolbuffers/protobuf/archive/v3.15.0.zip"],
)
load("@com_google_protobuf//:protobuf_deps.bzl", "protobuf_deps")
@@ -94,8 +94,13 @@ remote_apis_go_deps()
# Needed for the googleapis protos.
http_archive(
name = "googleapis",
- build_file = "BUILD.googleapis",
- sha256 = "7b6ea252f0b8fb5cd722f45feb83e115b689909bbb6a393a873b6cbad4ceae1d",
- strip_prefix = "googleapis-143084a2624b6591ee1f9d23e7f5241856642f4d",
- urls = ["https://github.com/googleapis/googleapis/archive/143084a2624b6591ee1f9d23e7f5241856642f4d.zip"],
+ sha256 = "b28c13e99001664eac5f1fb81b44d912d19fbc041e30772263251da131f6573c",
+ strip_prefix = "googleapis-bb964feba5980ed70c9fb8f84fe6e86694df65b0",
+ urls = ["https://github.com/googleapis/googleapis/archive/bb964feba5980ed70c9fb8f84fe6e86694df65b0.zip"],
+)
+
+load("@googleapis//:repository_rules.bzl", "switched_rules_by_language")
+
+switched_rules_by_language(
+ name = "com_google_googleapis_imports",
)
diff --git a/build/bazel/remote/asset/v1/BUILD b/build/bazel/remote/asset/v1/BUILD
index a09bea4..f360b4f 100644
--- a/build/bazel/remote/asset/v1/BUILD
+++ b/build/bazel/remote/asset/v1/BUILD
@@ -11,9 +11,9 @@ proto_library(
"//build/bazel/remote/execution/v2:remote_execution_proto",
"@com_google_protobuf//:duration_proto",
"@com_google_protobuf//:timestamp_proto",
- "@googleapis//:google_api_annotations_proto",
- "@googleapis//:google_api_http_proto",
- "@googleapis//:google_rpc_status_proto",
+ "@googleapis//google/api:annotations_proto",
+ "@googleapis//google/api:http_proto",
+ "@googleapis//google/rpc:status_proto",
],
)
diff --git a/build/bazel/remote/asset/v1/go/BUILD b/build/bazel/remote/asset/v1/go/BUILD
index 076422c..f1b1d26 100644
--- a/build/bazel/remote/asset/v1/go/BUILD
+++ b/build/bazel/remote/asset/v1/go/BUILD
@@ -10,8 +10,8 @@ go_proto_library(
proto = "//build/bazel/remote/asset/v1:remote_asset_proto",
deps = [
"//build/bazel/remote/execution/v2:remote_execution_go_proto",
- "@go_googleapis//google/api:annotations_go_proto",
- "@go_googleapis//google/rpc:status_go_proto",
+ "@org_golang_google_genproto//googleapis/api/annotations",
+ "@org_golang_google_genproto//googleapis/rpc/status",
],
)
diff --git a/build/bazel/remote/execution/v2/BUILD b/build/bazel/remote/execution/v2/BUILD
index 01f415b..e8edcf0 100644
--- a/build/bazel/remote/execution/v2/BUILD
+++ b/build/bazel/remote/execution/v2/BUILD
@@ -13,10 +13,10 @@ proto_library(
"@com_google_protobuf//:duration_proto",
"@com_google_protobuf//:timestamp_proto",
"@com_google_protobuf//:wrappers_proto",
- "@googleapis//:google_api_annotations_proto",
- "@googleapis//:google_api_http_proto",
- "@googleapis//:google_longrunning_operations_proto",
- "@googleapis//:google_rpc_status_proto",
+ "@googleapis//google/api:annotations_proto",
+ "@googleapis//google/api:http_proto",
+ "@googleapis//google/longrunning:operations_proto",
+ "@googleapis//google/rpc:status_proto",
],
)
diff --git a/build/bazel/remote/execution/v2/go/BUILD b/build/bazel/remote/execution/v2/go/BUILD
index 2f5dd2d..24cc7a0 100644
--- a/build/bazel/remote/execution/v2/go/BUILD
+++ b/build/bazel/remote/execution/v2/go/BUILD
@@ -10,9 +10,9 @@ go_proto_library(
proto = "//build/bazel/remote/execution/v2:remote_execution_proto",
deps = [
"//build/bazel/semver:semver_go_proto",
- "@go_googleapis//google/api:annotations_go_proto",
- "@go_googleapis//google/longrunning:longrunning_go_proto",
- "@go_googleapis//google/rpc:status_go_proto",
+ "@org_golang_google_genproto//googleapis/api/annotations",
+ "@org_golang_google_genproto//googleapis/longrunning",
+ "@org_golang_google_genproto//googleapis/rpc/status",
],
)
diff --git a/build/bazel/remote/execution/v2/remote_execution.pb.go b/build/bazel/remote/execution/v2/remote_execution.pb.go
index 58dc7f1..3e4feaa 100755
--- a/build/bazel/remote/execution/v2/remote_execution.pb.go
+++ b/build/bazel/remote/execution/v2/remote_execution.pb.go
@@ -14,8 +14,8 @@
// Code generated by protoc-gen-go. DO NOT EDIT.
// versions:
-// protoc-gen-go v1.25.0
-// protoc v3.10.1
+// protoc-gen-go v1.30.0
+// protoc v3.15.0
// source: build/bazel/remote/execution/v2/remote_execution.proto
package remoteexecution
@@ -23,11 +23,6 @@ package remoteexecution
import (
context "context"
semver "github.com/bazelbuild/remote-apis/build/bazel/semver"
- proto "github.com/golang/protobuf/proto"
- any "github.com/golang/protobuf/ptypes/any"
- duration "github.com/golang/protobuf/ptypes/duration"
- timestamp "github.com/golang/protobuf/ptypes/timestamp"
- wrappers "github.com/golang/protobuf/ptypes/wrappers"
_ "google.golang.org/genproto/googleapis/api/annotations"
longrunning "google.golang.org/genproto/googleapis/longrunning"
status "google.golang.org/genproto/googleapis/rpc/status"
@@ -36,6 +31,10 @@ import (
status1 "google.golang.org/grpc/status"
protoreflect "google.golang.org/protobuf/reflect/protoreflect"
protoimpl "google.golang.org/protobuf/runtime/protoimpl"
+ anypb "google.golang.org/protobuf/types/known/anypb"
+ durationpb "google.golang.org/protobuf/types/known/durationpb"
+ timestamppb "google.golang.org/protobuf/types/known/timestamppb"
+ wrapperspb "google.golang.org/protobuf/types/known/wrapperspb"
reflect "reflect"
sync "sync"
)
@@ -47,10 +46,6 @@ const (
_ = protoimpl.EnforceVersion(protoimpl.MaxVersion - 20)
)
-// This is a compile-time assertion that a sufficiently up-to-date version
-// of the legacy proto package is being used.
-const _ = proto.ProtoPackageIsVersion4
-
type ExecutionStage_Value int32
const (
@@ -146,44 +141,44 @@ const (
//
// SHA256TREE hashes are computed as follows:
//
- // - For blobs that are 1024 bytes or smaller, the hash is computed
- // using the regular SHA-256 digest function.
+ // - For blobs that are 1024 bytes or smaller, the hash is computed
+ // using the regular SHA-256 digest function.
//
- // - For blobs that are more than 1024 bytes in size, the hash is
- // computed as follows:
+ // - For blobs that are more than 1024 bytes in size, the hash is
+ // computed as follows:
//
- // 1. The blob is partitioned into a left (leading) and right
- // (trailing) blob. These blobs have lengths m and n
- // respectively, where m = 2^k and 0 < n <= m.
+ // 1. The blob is partitioned into a left (leading) and right
+ // (trailing) blob. These blobs have lengths m and n
+ // respectively, where m = 2^k and 0 < n <= m.
//
- // 2. Hashes of the left and right blob, Hash(left) and
- // Hash(right) respectively, are computed by recursively
- // applying the SHA256TREE algorithm.
+ // 2. Hashes of the left and right blob, Hash(left) and
+ // Hash(right) respectively, are computed by recursively
+ // applying the SHA256TREE algorithm.
//
- // 3. A single invocation is made to the SHA-256 block cipher with
- // the following parameters:
+ // 3. A single invocation is made to the SHA-256 block cipher with
+ // the following parameters:
//
- // M = Hash(left) || Hash(right)
- // H = {
- // 0xcbbb9d5d, 0x629a292a, 0x9159015a, 0x152fecd8,
- // 0x67332667, 0x8eb44a87, 0xdb0c2e0d, 0x47b5481d,
- // }
+ // M = Hash(left) || Hash(right)
+ // H = {
+ // 0xcbbb9d5d, 0x629a292a, 0x9159015a, 0x152fecd8,
+ // 0x67332667, 0x8eb44a87, 0xdb0c2e0d, 0x47b5481d,
+ // }
//
- // The values of H are the leading fractional parts of the
- // square roots of the 9th to the 16th prime number (23 to 53).
- // This differs from plain SHA-256, where the first eight prime
- // numbers (2 to 19) are used, thereby preventing trivial hash
- // collisions between small and large objects.
+ // The values of H are the leading fractional parts of the
+ // square roots of the 9th to the 16th prime number (23 to 53).
+ // This differs from plain SHA-256, where the first eight prime
+ // numbers (2 to 19) are used, thereby preventing trivial hash
+ // collisions between small and large objects.
//
- // 4. The hash of the full blob can then be obtained by
- // concatenating the outputs of the block cipher:
+ // 4. The hash of the full blob can then be obtained by
+ // concatenating the outputs of the block cipher:
//
- // Hash(blob) = a || b || c || d || e || f || g || h
+ // Hash(blob) = a || b || c || d || e || f || g || h
//
- // Addition of the original values of H, as normally done
- // through the use of the Davies-Meyer structure, is not
- // performed. This isn't necessary, as the block cipher is only
- // invoked once.
+ // Addition of the original values of H, as normally done
+ // through the use of the Davies-Meyer structure, is not
+ // performed. This isn't necessary, as the block cipher is only
+ // invoked once.
//
// Test vectors of this digest function can be found in the
// accompanying sha256tree_test_vectors.txt file.
@@ -428,7 +423,7 @@ type Action struct {
// timeout is too short. By encoding it directly in the `Action`, a lower
// timeout will result in a cache miss and the execution timeout will fail
// immediately, rather than whenever the cache entry gets evicted.
- Timeout *duration.Duration `protobuf:"bytes,6,opt,name=timeout,proto3" json:"timeout,omitempty"`
+ Timeout *durationpb.Duration `protobuf:"bytes,6,opt,name=timeout,proto3" json:"timeout,omitempty"`
// If true, then the `Action`'s result cannot be cached, and in-flight
// requests for the same `Action` may not be merged.
DoNotCache bool `protobuf:"varint,7,opt,name=do_not_cache,json=doNotCache,proto3" json:"do_not_cache,omitempty"`
@@ -496,7 +491,7 @@ func (x *Action) GetInputRootDigest() *Digest {
return nil
}
-func (x *Action) GetTimeout() *duration.Duration {
+func (x *Action) GetTimeout() *durationpb.Duration {
if x != nil {
return x.Timeout
}
@@ -585,7 +580,7 @@ type Command struct {
//
// DEPRECATED since v2.1: Use `output_paths` instead.
//
- // Deprecated: Do not use.
+ // Deprecated: Marked as deprecated in build/bazel/remote/execution/v2/remote_execution.proto.
OutputFiles []string `protobuf:"bytes,3,rep,name=output_files,json=outputFiles,proto3" json:"output_files,omitempty"`
// A list of the output directories that the client expects to retrieve from
// the action. Only the listed directories will be returned (an entire
@@ -617,7 +612,7 @@ type Command struct {
//
// DEPRECATED since 2.1: Use `output_paths` instead.
//
- // Deprecated: Do not use.
+ // Deprecated: Marked as deprecated in build/bazel/remote/execution/v2/remote_execution.proto.
OutputDirectories []string `protobuf:"bytes,4,rep,name=output_directories,json=outputDirectories,proto3" json:"output_directories,omitempty"`
// A list of the output paths that the client expects to retrieve from the
// action. Only the listed paths will be returned to the client as output.
@@ -657,7 +652,7 @@ type Command struct {
// the action. See documentation note in the
// [Action][build.bazel.remote.execution.v2.Action] for migration.
//
- // Deprecated: Do not use.
+ // Deprecated: Marked as deprecated in build/bazel/remote/execution/v2/remote_execution.proto.
Platform *Platform `protobuf:"bytes,5,opt,name=platform,proto3" json:"platform,omitempty"`
// The working directory, relative to the input root, for the command to run
// in. It must be a directory which exists in the input tree. If it is left
@@ -723,7 +718,7 @@ func (x *Command) GetEnvironmentVariables() []*Command_EnvironmentVariable {
return nil
}
-// Deprecated: Do not use.
+// Deprecated: Marked as deprecated in build/bazel/remote/execution/v2/remote_execution.proto.
func (x *Command) GetOutputFiles() []string {
if x != nil {
return x.OutputFiles
@@ -731,7 +726,7 @@ func (x *Command) GetOutputFiles() []string {
return nil
}
-// Deprecated: Do not use.
+// Deprecated: Marked as deprecated in build/bazel/remote/execution/v2/remote_execution.proto.
func (x *Command) GetOutputDirectories() []string {
if x != nil {
return x.OutputDirectories
@@ -746,7 +741,7 @@ func (x *Command) GetOutputPaths() []string {
return nil
}
-// Deprecated: Do not use.
+// Deprecated: Marked as deprecated in build/bazel/remote/execution/v2/remote_execution.proto.
func (x *Command) GetPlatform() *Platform {
if x != nil {
return x.Platform
@@ -836,20 +831,20 @@ func (x *Platform) GetProperties() []*Platform_Property {
// value, the following restrictions MUST be obeyed when constructing a
// a `Directory`:
//
-// * Every child in the directory must have a path of exactly one segment.
-// Multiple levels of directory hierarchy may not be collapsed.
-// * Each child in the directory must have a unique path segment (file name).
-// Note that while the API itself is case-sensitive, the environment where
-// the Action is executed may or may not be case-sensitive. That is, it is
-// legal to call the API with a Directory that has both "Foo" and "foo" as
-// children, but the Action may be rejected by the remote system upon
-// execution.
-// * The files, directories and symlinks in the directory must each be sorted
-// in lexicographical order by path. The path strings must be sorted by code
-// point, equivalently, by UTF-8 bytes.
-// * The [NodeProperties][build.bazel.remote.execution.v2.NodeProperty] of files,
-// directories, and symlinks must be sorted in lexicographical order by
-// property name.
+// - Every child in the directory must have a path of exactly one segment.
+// Multiple levels of directory hierarchy may not be collapsed.
+// - Each child in the directory must have a unique path segment (file name).
+// Note that while the API itself is case-sensitive, the environment where
+// the Action is executed may or may not be case-sensitive. That is, it is
+// legal to call the API with a Directory that has both "Foo" and "foo" as
+// children, but the Action may be rejected by the remote system upon
+// execution.
+// - The files, directories and symlinks in the directory must each be sorted
+// in lexicographical order by path. The path strings must be sorted by code
+// point, equivalently, by UTF-8 bytes.
+// - The [NodeProperties][build.bazel.remote.execution.v2.NodeProperty] of files,
+// directories, and symlinks must be sorted in lexicographical order by
+// property name.
//
// A `Directory` that obeys the restrictions is said to be in canonical form.
//
@@ -859,46 +854,49 @@ func (x *Platform) GetProperties() []*Platform_Property {
//
// ```json
// // (Directory proto)
-// {
-// files: [
-// {
-// name: "bar",
-// digest: {
-// hash: "4a73bc9d03...",
-// size: 65534
-// },
-// node_properties: [
-// {
-// "name": "MTime",
-// "value": "2017-01-15T01:30:15.01Z"
-// }
-// ]
-// }
-// ],
-// directories: [
-// {
-// name: "foo",
-// digest: {
-// hash: "4cf2eda940...",
-// size: 43
-// }
-// }
-// ]
-// }
+//
+// {
+// files: [
+// {
+// name: "bar",
+// digest: {
+// hash: "4a73bc9d03...",
+// size: 65534
+// },
+// node_properties: [
+// {
+// "name": "MTime",
+// "value": "2017-01-15T01:30:15.01Z"
+// }
+// ]
+// }
+// ],
+// directories: [
+// {
+// name: "foo",
+// digest: {
+// hash: "4cf2eda940...",
+// size: 43
+// }
+// }
+// ]
+// }
//
// // (Directory proto with hash "4cf2eda940..." and size 43)
-// {
-// files: [
-// {
-// name: "baz",
-// digest: {
-// hash: "b2c941073e...",
-// size: 1294,
-// },
-// is_executable: true
-// }
-// ]
-// }
+//
+// {
+// files: [
+// {
+// name: "baz",
+// digest: {
+// hash: "b2c941073e...",
+// size: 1294,
+// },
+// is_executable: true
+// }
+// ]
+// }
+//
// ```
type Directory struct {
state protoimpl.MessageState
@@ -1040,7 +1038,6 @@ func (x *NodeProperty) GetValue() string {
// [DirectoryNodes][build.bazel.remote.execution.v2.DirectoryNode], and
// [SymlinkNodes][build.bazel.remote.execution.v2.SymlinkNode]. The server is
// responsible for specifying the properties that it accepts.
-//
type NodeProperties struct {
state protoimpl.MessageState
sizeCache protoimpl.SizeCache
@@ -1050,9 +1047,9 @@ type NodeProperties struct {
// [NodeProperties][build.bazel.remote.execution.v2.NodeProperty].
Properties []*NodeProperty `protobuf:"bytes,1,rep,name=properties,proto3" json:"properties,omitempty"`
// The file's last modification timestamp.
- Mtime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=mtime,proto3" json:"mtime,omitempty"`
+ Mtime *timestamppb.Timestamp `protobuf:"bytes,2,opt,name=mtime,proto3" json:"mtime,omitempty"`
// The UNIX file mode, e.g., 0755.
- UnixMode *wrappers.UInt32Value `protobuf:"bytes,3,opt,name=unix_mode,json=unixMode,proto3" json:"unix_mode,omitempty"`
+ UnixMode *wrapperspb.UInt32Value `protobuf:"bytes,3,opt,name=unix_mode,json=unixMode,proto3" json:"unix_mode,omitempty"`
}
func (x *NodeProperties) Reset() {
@@ -1094,14 +1091,14 @@ func (x *NodeProperties) GetProperties() []*NodeProperty {
return nil
}
-func (x *NodeProperties) GetMtime() *timestamp.Timestamp {
+func (x *NodeProperties) GetMtime() *timestamppb.Timestamp {
if x != nil {
return x.Mtime
}
return nil
}
-func (x *NodeProperties) GetUnixMode() *wrappers.UInt32Value {
+func (x *NodeProperties) GetUnixMode() *wrapperspb.UInt32Value {
if x != nil {
return x.UnixMode
}
@@ -1418,19 +1415,19 @@ type ExecutedActionMetadata struct {
// The name of the worker which ran the execution.
Worker string `protobuf:"bytes,1,opt,name=worker,proto3" json:"worker,omitempty"`
// When was the action added to the queue.
- QueuedTimestamp *timestamp.Timestamp `protobuf:"bytes,2,opt,name=queued_timestamp,json=queuedTimestamp,proto3" json:"queued_timestamp,omitempty"`
+ QueuedTimestamp *timestamppb.Timestamp `protobuf:"bytes,2,opt,name=queued_timestamp,json=queuedTimestamp,proto3" json:"queued_timestamp,omitempty"`
// When the worker received the action.
- WorkerStartTimestamp *timestamp.Timestamp `protobuf:"bytes,3,opt,name=worker_start_timestamp,json=workerStartTimestamp,proto3" json:"worker_start_timestamp,omitempty"`
+ WorkerStartTimestamp *timestamppb.Timestamp `protobuf:"bytes,3,opt,name=worker_start_timestamp,json=workerStartTimestamp,proto3" json:"worker_start_timestamp,omitempty"`
// When the worker completed the action, including all stages.
- WorkerCompletedTimestamp *timestamp.Timestamp `protobuf:"bytes,4,opt,name=worker_completed_timestamp,json=workerCompletedTimestamp,proto3" json:"worker_completed_timestamp,omitempty"`
+ WorkerCompletedTimestamp *timestamppb.Timestamp `protobuf:"bytes,4,opt,name=worker_completed_timestamp,json=workerCompletedTimestamp,proto3" json:"worker_completed_timestamp,omitempty"`
// When the worker started fetching action inputs.
- InputFetchStartTimestamp *timestamp.Timestamp `protobuf:"bytes,5,opt,name=input_fetch_start_timestamp,json=inputFetchStartTimestamp,proto3" json:"input_fetch_start_timestamp,omitempty"`
+ InputFetchStartTimestamp *timestamppb.Timestamp `protobuf:"bytes,5,opt,name=input_fetch_start_timestamp,json=inputFetchStartTimestamp,proto3" json:"input_fetch_start_timestamp,omitempty"`
// When the worker finished fetching action inputs.
- InputFetchCompletedTimestamp *timestamp.Timestamp `protobuf:"bytes,6,opt,name=input_fetch_completed_timestamp,json=inputFetchCompletedTimestamp,proto3" json:"input_fetch_completed_timestamp,omitempty"`
+ InputFetchCompletedTimestamp *timestamppb.Timestamp `protobuf:"bytes,6,opt,name=input_fetch_completed_timestamp,json=inputFetchCompletedTimestamp,proto3" json:"input_fetch_completed_timestamp,omitempty"`
// When the worker started executing the action command.
- ExecutionStartTimestamp *timestamp.Timestamp `protobuf:"bytes,7,opt,name=execution_start_timestamp,json=executionStartTimestamp,proto3" json:"execution_start_timestamp,omitempty"`
+ ExecutionStartTimestamp *timestamppb.Timestamp `protobuf:"bytes,7,opt,name=execution_start_timestamp,json=executionStartTimestamp,proto3" json:"execution_start_timestamp,omitempty"`
// When the worker completed executing the action command.
- ExecutionCompletedTimestamp *timestamp.Timestamp `protobuf:"bytes,8,opt,name=execution_completed_timestamp,json=executionCompletedTimestamp,proto3" json:"execution_completed_timestamp,omitempty"`
+ ExecutionCompletedTimestamp *timestamppb.Timestamp `protobuf:"bytes,8,opt,name=execution_completed_timestamp,json=executionCompletedTimestamp,proto3" json:"execution_completed_timestamp,omitempty"`
// New in v2.3: the amount of time the worker spent executing the action
// command, potentially computed using a worker-specific virtual clock.
//
@@ -1448,15 +1445,15 @@ type ExecutedActionMetadata struct {
// [Action][[build.bazel.remote.execution.v2.Action]'s `timeout`. There is no
// relationship between the virtual execution duration and the values of
// `execution_start_timestamp` and `execution_completed_timestamp`.
- VirtualExecutionDuration *duration.Duration `protobuf:"bytes,12,opt,name=virtual_execution_duration,json=virtualExecutionDuration,proto3" json:"virtual_execution_duration,omitempty"`
+ VirtualExecutionDuration *durationpb.Duration `protobuf:"bytes,12,opt,name=virtual_execution_duration,json=virtualExecutionDuration,proto3" json:"virtual_execution_duration,omitempty"`
// When the worker started uploading action outputs.
- OutputUploadStartTimestamp *timestamp.Timestamp `protobuf:"bytes,9,opt,name=output_upload_start_timestamp,json=outputUploadStartTimestamp,proto3" json:"output_upload_start_timestamp,omitempty"`
+ OutputUploadStartTimestamp *timestamppb.Timestamp `protobuf:"bytes,9,opt,name=output_upload_start_timestamp,json=outputUploadStartTimestamp,proto3" json:"output_upload_start_timestamp,omitempty"`
// When the worker finished uploading action outputs.
- OutputUploadCompletedTimestamp *timestamp.Timestamp `protobuf:"bytes,10,opt,name=output_upload_completed_timestamp,json=outputUploadCompletedTimestamp,proto3" json:"output_upload_completed_timestamp,omitempty"`
+ OutputUploadCompletedTimestamp *timestamppb.Timestamp `protobuf:"bytes,10,opt,name=output_upload_completed_timestamp,json=outputUploadCompletedTimestamp,proto3" json:"output_upload_completed_timestamp,omitempty"`
// Details that are specific to the kind of worker used. For example,
// on POSIX-like systems this could contain a message with
// getrusage(2) statistics.
- AuxiliaryMetadata []*any.Any `protobuf:"bytes,11,rep,name=auxiliary_metadata,json=auxiliaryMetadata,proto3" json:"auxiliary_metadata,omitempty"`
+ AuxiliaryMetadata []*anypb.Any `protobuf:"bytes,11,rep,name=auxiliary_metadata,json=auxiliaryMetadata,proto3" json:"auxiliary_metadata,omitempty"`
}
func (x *ExecutedActionMetadata) Reset() {
@@ -1498,77 +1495,77 @@ func (x *ExecutedActionMetadata) GetWorker() string {
return ""
}
-func (x *ExecutedActionMetadata) GetQueuedTimestamp() *timestamp.Timestamp {
+func (x *ExecutedActionMetadata) GetQueuedTimestamp() *timestamppb.Timestamp {
if x != nil {
return x.QueuedTimestamp
}
return nil
}
-func (x *ExecutedActionMetadata) GetWorkerStartTimestamp() *timestamp.Timestamp {
+func (x *ExecutedActionMetadata) GetWorkerStartTimestamp() *timestamppb.Timestamp {
if x != nil {
return x.WorkerStartTimestamp
}
return nil
}
-func (x *ExecutedActionMetadata) GetWorkerCompletedTimestamp() *timestamp.Timestamp {
+func (x *ExecutedActionMetadata) GetWorkerCompletedTimestamp() *timestamppb.Timestamp {
if x != nil {
return x.WorkerCompletedTimestamp
}
return nil
}
-func (x *ExecutedActionMetadata) GetInputFetchStartTimestamp() *timestamp.Timestamp {
+func (x *ExecutedActionMetadata) GetInputFetchStartTimestamp() *timestamppb.Timestamp {
if x != nil {
return x.InputFetchStartTimestamp
}
return nil
}
-func (x *ExecutedActionMetadata) GetInputFetchCompletedTimestamp() *timestamp.Timestamp {
+func (x *ExecutedActionMetadata) GetInputFetchCompletedTimestamp() *timestamppb.Timestamp {
if x != nil {
return x.InputFetchCompletedTimestamp
}
return nil
}
-func (x *ExecutedActionMetadata) GetExecutionStartTimestamp() *timestamp.Timestamp {
+func (x *ExecutedActionMetadata) GetExecutionStartTimestamp() *timestamppb.Timestamp {
if x != nil {
return x.ExecutionStartTimestamp
}
return nil
}
-func (x *ExecutedActionMetadata) GetExecutionCompletedTimestamp() *timestamp.Timestamp {
+func (x *ExecutedActionMetadata) GetExecutionCompletedTimestamp() *timestamppb.Timestamp {
if x != nil {
return x.ExecutionCompletedTimestamp
}
return nil
}
-func (x *ExecutedActionMetadata) GetVirtualExecutionDuration() *duration.Duration {
+func (x *ExecutedActionMetadata) GetVirtualExecutionDuration() *durationpb.Duration {
if x != nil {
return x.VirtualExecutionDuration
}
return nil
}
-func (x *ExecutedActionMetadata) GetOutputUploadStartTimestamp() *timestamp.Timestamp {
+func (x *ExecutedActionMetadata) GetOutputUploadStartTimestamp() *timestamppb.Timestamp {
if x != nil {
return x.OutputUploadStartTimestamp
}
return nil
}
-func (x *ExecutedActionMetadata) GetOutputUploadCompletedTimestamp() *timestamp.Timestamp {
+func (x *ExecutedActionMetadata) GetOutputUploadCompletedTimestamp() *timestamppb.Timestamp {
if x != nil {
return x.OutputUploadCompletedTimestamp
}
return nil
}
-func (x *ExecutedActionMetadata) GetAuxiliaryMetadata() []*any.Any {
+func (x *ExecutedActionMetadata) GetAuxiliaryMetadata() []*anypb.Any {
if x != nil {
return x.AuxiliaryMetadata
}
@@ -1618,7 +1615,7 @@ type ActionResult struct {
// DEPRECATED as of v2.1. Servers that wish to be compatible with v2.0 API
// should still populate this field in addition to `output_symlinks`.
//
- // Deprecated: Do not use.
+ // Deprecated: Marked as deprecated in build/bazel/remote/execution/v2/remote_execution.proto.
OutputFileSymlinks []*OutputSymlink `protobuf:"bytes,10,rep,name=output_file_symlinks,json=outputFileSymlinks,proto3" json:"output_file_symlinks,omitempty"`
// New in v2.1: this field will only be populated if the command
// `output_paths` field was used, and not the pre v2.1 `output_files` or
@@ -1650,49 +1647,53 @@ type ActionResult struct {
//
// ```json
// // OutputDirectory proto:
- // {
- // path: "a/b/dir"
- // tree_digest: {
- // hash: "4a73bc9d03...",
- // size: 55
- // }
- // }
+ //
+ // {
+ // path: "a/b/dir"
+ // tree_digest: {
+ // hash: "4a73bc9d03...",
+ // size: 55
+ // }
+ // }
+ //
// // Tree proto with hash "4a73bc9d03..." and size 55:
- // {
- // root: {
- // files: [
- // {
- // name: "bar",
- // digest: {
- // hash: "4a73bc9d03...",
- // size: 65534
- // }
- // }
- // ],
- // directories: [
- // {
- // name: "foo",
- // digest: {
- // hash: "4cf2eda940...",
- // size: 43
- // }
- // }
- // ]
- // }
- // children : {
- // // (Directory proto with hash "4cf2eda940..." and size 43)
- // files: [
- // {
- // name: "baz",
- // digest: {
- // hash: "b2c941073e...",
- // size: 1294,
- // },
- // is_executable: true
- // }
- // ]
- // }
- // }
+ //
+ // {
+ // root: {
+ // files: [
+ // {
+ // name: "bar",
+ // digest: {
+ // hash: "4a73bc9d03...",
+ // size: 65534
+ // }
+ // }
+ // ],
+ // directories: [
+ // {
+ // name: "foo",
+ // digest: {
+ // hash: "4cf2eda940...",
+ // size: 43
+ // }
+ // }
+ // ]
+ // }
+ // children : {
+ // // (Directory proto with hash "4cf2eda940..." and size 43)
+ // files: [
+ // {
+ // name: "baz",
+ // digest: {
+ // hash: "b2c941073e...",
+ // size: 1294,
+ // },
+ // is_executable: true
+ // }
+ // ]
+ // }
+ // }
+ //
// ```
// If an output of the same name as listed in `output_files` of
// the Command was found in `output_directories`, but was not a directory, the
@@ -1717,7 +1718,7 @@ type ActionResult struct {
// DEPRECATED as of v2.1. Servers that wish to be compatible with v2.0 API
// should still populate this field in addition to `output_symlinks`.
//
- // Deprecated: Do not use.
+ // Deprecated: Marked as deprecated in build/bazel/remote/execution/v2/remote_execution.proto.
OutputDirectorySymlinks []*OutputSymlink `protobuf:"bytes,11,rep,name=output_directory_symlinks,json=outputDirectorySymlinks,proto3" json:"output_directory_symlinks,omitempty"`
// The exit code of the command.
ExitCode int32 `protobuf:"varint,4,opt,name=exit_code,json=exitCode,proto3" json:"exit_code,omitempty"`
@@ -1786,7 +1787,7 @@ func (x *ActionResult) GetOutputFiles() []*OutputFile {
return nil
}
-// Deprecated: Do not use.
+// Deprecated: Marked as deprecated in build/bazel/remote/execution/v2/remote_execution.proto.
func (x *ActionResult) GetOutputFileSymlinks() []*OutputSymlink {
if x != nil {
return x.OutputFileSymlinks
@@ -1808,7 +1809,7 @@ func (x *ActionResult) GetOutputDirectories() []*OutputDirectory {
return nil
}
-// Deprecated: Do not use.
+// Deprecated: Marked as deprecated in build/bazel/remote/execution/v2/remote_execution.proto.
func (x *ActionResult) GetOutputDirectorySymlinks() []*OutputSymlink {
if x != nil {
return x.OutputDirectorySymlinks
@@ -2039,13 +2040,13 @@ type OutputDirectory struct {
// instantiated on a local file system by scanning through it
// sequentially:
//
- // - All directories with the same binary representation are stored
- // exactly once.
- // - All directories, apart from the root directory, are referenced by
- // at least one parent directory.
- // - Directories are stored in topological order, with parents being
- // stored before the child. The root directory is thus the first to
- // be stored.
+ // - All directories with the same binary representation are stored
+ // exactly once.
+ // - All directories, apart from the root directory, are referenced by
+ // at least one parent directory.
+ // - Directories are stored in topological order, with parents being
+ // stored before the child. The root directory is thus the first to
+ // be stored.
//
// Additionally, the Tree MUST be encoded as a stream of records,
// where each record has the following format:
@@ -2053,9 +2054,9 @@ type OutputDirectory struct {
// - A tag byte, having one of the following two values:
// - (1 << 3) | 2 == 0x0a: First record (the root directory).
// - (2 << 3) | 2 == 0x12: Any subsequent records (child directories).
- // - The size of the directory, encoded as a base 128 varint.
- // - The contents of the directory, encoded as a binary serialized
- // Protobuf message.
+ // - The size of the directory, encoded as a base 128 varint.
+ // - The contents of the directory, encoded as a binary serialized
+ // Protobuf message.
//
// This encoding is a subset of the Protobuf wire format of the Tree
// message. As it is only permitted to store data associated with
@@ -5521,19 +5522,19 @@ var file_build_bazel_remote_execution_v2_remote_execution_proto_rawDesc = []byte
0x32, 0x2e, 0x45, 0x78, 0x65, 0x63, 0x75, 0x74, 0x65, 0x52, 0x65, 0x71, 0x75, 0x65, 0x73, 0x74,
0x1a, 0x1d, 0x2e, 0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x6c, 0x6f, 0x6e, 0x67, 0x72, 0x75,
0x6e, 0x6e, 0x69, 0x6e, 0x67, 0x2e, 0x4f, 0x70, 0x65, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x22,
- 0x31, 0x82, 0xd3, 0xe4, 0x93, 0x02, 0x2b, 0x22, 0x26, 0x2f, 0x76, 0x32, 0x2f, 0x7b, 0x69, 0x6e,
- 0x73, 0x74, 0x61, 0x6e, 0x63, 0x65, 0x5f, 0x6e, 0x61, 0x6d, 0x65, 0x3d, 0x2a, 0x2a, 0x7d, 0x2f,
- 0x61, 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x3a, 0x65, 0x78, 0x65, 0x63, 0x75, 0x74, 0x65, 0x3a,
- 0x01, 0x2a, 0x30, 0x01, 0x12, 0x9a, 0x01, 0x0a, 0x0d, 0x57, 0x61, 0x69, 0x74, 0x45, 0x78, 0x65,
+ 0x31, 0x82, 0xd3, 0xe4, 0x93, 0x02, 0x2b, 0x3a, 0x01, 0x2a, 0x22, 0x26, 0x2f, 0x76, 0x32, 0x2f,
+ 0x7b, 0x69, 0x6e, 0x73, 0x74, 0x61, 0x6e, 0x63, 0x65, 0x5f, 0x6e, 0x61, 0x6d, 0x65, 0x3d, 0x2a,
+ 0x2a, 0x7d, 0x2f, 0x61, 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x3a, 0x65, 0x78, 0x65, 0x63, 0x75,
+ 0x74, 0x65, 0x30, 0x01, 0x12, 0x9a, 0x01, 0x0a, 0x0d, 0x57, 0x61, 0x69, 0x74, 0x45, 0x78, 0x65,
0x63, 0x75, 0x74, 0x69, 0x6f, 0x6e, 0x12, 0x35, 0x2e, 0x62, 0x75, 0x69, 0x6c, 0x64, 0x2e, 0x62,
0x61, 0x7a, 0x65, 0x6c, 0x2e, 0x72, 0x65, 0x6d, 0x6f, 0x74, 0x65, 0x2e, 0x65, 0x78, 0x65, 0x63,
0x75, 0x74, 0x69, 0x6f, 0x6e, 0x2e, 0x76, 0x32, 0x2e, 0x57, 0x61, 0x69, 0x74, 0x45, 0x78, 0x65,
0x63, 0x75, 0x74, 0x69, 0x6f, 0x6e, 0x52, 0x65, 0x71, 0x75, 0x65, 0x73, 0x74, 0x1a, 0x1d, 0x2e,
0x67, 0x6f, 0x6f, 0x67, 0x6c, 0x65, 0x2e, 0x6c, 0x6f, 0x6e, 0x67, 0x72, 0x75, 0x6e, 0x6e, 0x69,
0x6e, 0x67, 0x2e, 0x4f, 0x70, 0x65, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x22, 0x31, 0x82, 0xd3,
- 0xe4, 0x93, 0x02, 0x2b, 0x22, 0x26, 0x2f, 0x76, 0x32, 0x2f, 0x7b, 0x6e, 0x61, 0x6d, 0x65, 0x3d,
- 0x6f, 0x70, 0x65, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x2f, 0x2a, 0x2a, 0x7d, 0x3a, 0x77,
- 0x61, 0x69, 0x74, 0x45, 0x78, 0x65, 0x63, 0x75, 0x74, 0x69, 0x6f, 0x6e, 0x3a, 0x01, 0x2a, 0x30,
+ 0xe4, 0x93, 0x02, 0x2b, 0x3a, 0x01, 0x2a, 0x22, 0x26, 0x2f, 0x76, 0x32, 0x2f, 0x7b, 0x6e, 0x61,
+ 0x6d, 0x65, 0x3d, 0x6f, 0x70, 0x65, 0x72, 0x61, 0x74, 0x69, 0x6f, 0x6e, 0x73, 0x2f, 0x2a, 0x2a,
+ 0x7d, 0x3a, 0x77, 0x61, 0x69, 0x74, 0x45, 0x78, 0x65, 0x63, 0x75, 0x74, 0x69, 0x6f, 0x6e, 0x30,
0x01, 0x32, 0xd6, 0x03, 0x0a, 0x0b, 0x41, 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x43, 0x61, 0x63, 0x68,
0x65, 0x12, 0xd7, 0x01, 0x0a, 0x0f, 0x47, 0x65, 0x74, 0x41, 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x52,
0x65, 0x73, 0x75, 0x6c, 0x74, 0x12, 0x37, 0x2e, 0x62, 0x75, 0x69, 0x6c, 0x64, 0x2e, 0x62, 0x61,
@@ -5557,13 +5558,13 @@ var file_build_bazel_remote_execution_v2_remote_execution_proto_rawDesc = []byte
0x2e, 0x62, 0x75, 0x69, 0x6c, 0x64, 0x2e, 0x62, 0x61, 0x7a, 0x65, 0x6c, 0x2e, 0x72, 0x65, 0x6d,
0x6f, 0x74, 0x65, 0x2e, 0x65, 0x78, 0x65, 0x63, 0x75, 0x74, 0x69, 0x6f, 0x6e, 0x2e, 0x76, 0x32,
0x2e, 0x41, 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x52, 0x65, 0x73, 0x75, 0x6c, 0x74, 0x22, 0x6b, 0x82,
- 0xd3, 0xe4, 0x93, 0x02, 0x65, 0x1a, 0x54, 0x2f, 0x76, 0x32, 0x2f, 0x7b, 0x69, 0x6e, 0x73, 0x74,
- 0x61, 0x6e, 0x63, 0x65, 0x5f, 0x6e, 0x61, 0x6d, 0x65, 0x3d, 0x2a, 0x2a, 0x7d, 0x2f, 0x61, 0x63,
- 0x74, 0x69, 0x6f, 0x6e, 0x52, 0x65, 0x73, 0x75, 0x6c, 0x74, 0x73, 0x2f, 0x7b, 0x61, 0x63, 0x74,
- 0x69, 0x6f, 0x6e, 0x5f, 0x64, 0x69, 0x67, 0x65, 0x73, 0x74, 0x2e, 0x68, 0x61, 0x73, 0x68, 0x7d,
- 0x2f, 0x7b, 0x61, 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x5f, 0x64, 0x69, 0x67, 0x65, 0x73, 0x74, 0x2e,
- 0x73, 0x69, 0x7a, 0x65, 0x5f, 0x62, 0x79, 0x74, 0x65, 0x73, 0x7d, 0x3a, 0x0d, 0x61, 0x63, 0x74,
- 0x69, 0x6f, 0x6e, 0x5f, 0x72, 0x65, 0x73, 0x75, 0x6c, 0x74, 0x32, 0x9b, 0x06, 0x0a, 0x19, 0x43,
+ 0xd3, 0xe4, 0x93, 0x02, 0x65, 0x3a, 0x0d, 0x61, 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x5f, 0x72, 0x65,
+ 0x73, 0x75, 0x6c, 0x74, 0x1a, 0x54, 0x2f, 0x76, 0x32, 0x2f, 0x7b, 0x69, 0x6e, 0x73, 0x74, 0x61,
+ 0x6e, 0x63, 0x65, 0x5f, 0x6e, 0x61, 0x6d, 0x65, 0x3d, 0x2a, 0x2a, 0x7d, 0x2f, 0x61, 0x63, 0x74,
+ 0x69, 0x6f, 0x6e, 0x52, 0x65, 0x73, 0x75, 0x6c, 0x74, 0x73, 0x2f, 0x7b, 0x61, 0x63, 0x74, 0x69,
+ 0x6f, 0x6e, 0x5f, 0x64, 0x69, 0x67, 0x65, 0x73, 0x74, 0x2e, 0x68, 0x61, 0x73, 0x68, 0x7d, 0x2f,
+ 0x7b, 0x61, 0x63, 0x74, 0x69, 0x6f, 0x6e, 0x5f, 0x64, 0x69, 0x67, 0x65, 0x73, 0x74, 0x2e, 0x73,
+ 0x69, 0x7a, 0x65, 0x5f, 0x62, 0x79, 0x74, 0x65, 0x73, 0x7d, 0x32, 0x9b, 0x06, 0x0a, 0x19, 0x43,
0x6f, 0x6e, 0x74, 0x65, 0x6e, 0x74, 0x41, 0x64, 0x64, 0x72, 0x65, 0x73, 0x73, 0x61, 0x62, 0x6c,
0x65, 0x53, 0x74, 0x6f, 0x72, 0x61, 0x67, 0x65, 0x12, 0xbc, 0x01, 0x0a, 0x10, 0x46, 0x69, 0x6e,
0x64, 0x4d, 0x69, 0x73, 0x73, 0x69, 0x6e, 0x67, 0x42, 0x6c, 0x6f, 0x62, 0x73, 0x12, 0x38, 0x2e,
@@ -5574,10 +5575,10 @@ var file_build_bazel_remote_execution_v2_remote_execution_proto_rawDesc = []byte
0x62, 0x61, 0x7a, 0x65, 0x6c, 0x2e, 0x72, 0x65, 0x6d, 0x6f, 0x74, 0x65, 0x2e, 0x65, 0x78, 0x65,
0x63, 0x75, 0x74, 0x69, 0x6f, 0x6e, 0x2e, 0x76, 0x32, 0x2e, 0x46, 0x69, 0x6e, 0x64, 0x4d, 0x69,
0x73, 0x73, 0x69, 0x6e, 0x67, 0x42, 0x6c, 0x6f, 0x62, 0x73, 0x52, 0x65, 0x73, 0x70, 0x6f, 0x6e,
- 0x73, 0x65, 0x22, 0x33, 0x82, 0xd3, 0xe4, 0x93, 0x02, 0x2d, 0x22, 0x28, 0x2f, 0x76, 0x32, 0x2f,
- 0x7b, 0x69, 0x6e, 0x73, 0x74, 0x61, 0x6e, 0x63, 0x65, 0x5f, 0x6e, 0x61, 0x6d, 0x65, 0x3d, 0x2a,
- 0x2a, 0x7d, 0x2f, 0x62, 0x6c, 0x6f, 0x62, 0x73, 0x3a, 0x66, 0x69, 0x6e, 0x64, 0x4d, 0x69, 0x73,
- 0x73, 0x69, 0x6e, 0x67, 0x3a, 0x01, 0x2a, 0x12, 0xbc, 0x01, 0x0a, 0x10, 0x42, 0x61, 0x74, 0x63,
+ 0x73, 0x65, 0x22, 0x33, 0x82, 0xd3, 0xe4, 0x93, 0x02, 0x2d, 0x3a, 0x01, 0x2a, 0x22, 0x28, 0x2f,
+ 0x76, 0x32, 0x2f, 0x7b, 0x69, 0x6e, 0x73, 0x74, 0x61, 0x6e, 0x63, 0x65, 0x5f, 0x6e, 0x61, 0x6d,
+ 0x65, 0x3d, 0x2a, 0x2a, 0x7d, 0x2f, 0x62, 0x6c, 0x6f, 0x62, 0x73, 0x3a, 0x66, 0x69, 0x6e, 0x64,
+ 0x4d, 0x69, 0x73, 0x73, 0x69, 0x6e, 0x67, 0x12, 0xbc, 0x01, 0x0a, 0x10, 0x42, 0x61, 0x74, 0x63,
0x68, 0x55, 0x70, 0x64, 0x61, 0x74, 0x65, 0x42, 0x6c, 0x6f, 0x62, 0x73, 0x12, 0x38, 0x2e, 0x62,
0x75, 0x69, 0x6c, 0x64, 0x2e, 0x62, 0x61, 0x7a, 0x65, 0x6c, 0x2e, 0x72, 0x65, 0x6d, 0x6f, 0x74,
0x65, 0x2e, 0x65, 0x78, 0x65, 0x63, 0x75, 0x74, 0x69, 0x6f, 0x6e, 0x2e, 0x76, 0x32, 0x2e, 0x42,
@@ -5586,10 +5587,10 @@ var file_build_bazel_remote_execution_v2_remote_execution_proto_rawDesc = []byte
0x61, 0x7a, 0x65, 0x6c, 0x2e, 0x72, 0x65, 0x6d, 0x6f, 0x74, 0x65, 0x2e, 0x65, 0x78, 0x65, 0x63,
0x75, 0x74, 0x69, 0x6f, 0x6e, 0x2e, 0x76, 0x32, 0x2e, 0x42, 0x61, 0x74, 0x63, 0x68, 0x55, 0x70,
0x64, 0x61, 0x74, 0x65, 0x42, 0x6c, 0x6f, 0x62, 0x73, 0x52, 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73,
- 0x65, 0x22, 0x33, 0x82, 0xd3, 0xe4, 0x93, 0x02, 0x2d, 0x22, 0x28, 0x2f, 0x76, 0x32, 0x2f, 0x7b,
- 0x69, 0x6e, 0x73, 0x74, 0x61, 0x6e, 0x63, 0x65, 0x5f, 0x6e, 0x61, 0x6d, 0x65, 0x3d, 0x2a, 0x2a,
- 0x7d, 0x2f, 0x62, 0x6c, 0x6f, 0x62, 0x73, 0x3a, 0x62, 0x61, 0x74, 0x63, 0x68, 0x55, 0x70, 0x64,
- 0x61, 0x74, 0x65, 0x3a, 0x01, 0x2a, 0x12, 0xb4, 0x01, 0x0a, 0x0e, 0x42, 0x61, 0x74, 0x63, 0x68,
+ 0x65, 0x22, 0x33, 0x82, 0xd3, 0xe4, 0x93, 0x02, 0x2d, 0x3a, 0x01, 0x2a, 0x22, 0x28, 0x2f, 0x76,
+ 0x32, 0x2f, 0x7b, 0x69, 0x6e, 0x73, 0x74, 0x61, 0x6e, 0x63, 0x65, 0x5f, 0x6e, 0x61, 0x6d, 0x65,
+ 0x3d, 0x2a, 0x2a, 0x7d, 0x2f, 0x62, 0x6c, 0x6f, 0x62, 0x73, 0x3a, 0x62, 0x61, 0x74, 0x63, 0x68,
+ 0x55, 0x70, 0x64, 0x61, 0x74, 0x65, 0x12, 0xb4, 0x01, 0x0a, 0x0e, 0x42, 0x61, 0x74, 0x63, 0x68,
0x52, 0x65, 0x61, 0x64, 0x42, 0x6c, 0x6f, 0x62, 0x73, 0x12, 0x36, 0x2e, 0x62, 0x75, 0x69, 0x6c,
0x64, 0x2e, 0x62, 0x61, 0x7a, 0x65, 0x6c, 0x2e, 0x72, 0x65, 0x6d, 0x6f, 0x74, 0x65, 0x2e, 0x65,
0x78, 0x65, 0x63, 0x75, 0x74, 0x69, 0x6f, 0x6e, 0x2e, 0x76, 0x32, 0x2e, 0x42, 0x61, 0x74, 0x63,
@@ -5598,9 +5599,9 @@ var file_build_bazel_remote_execution_v2_remote_execution_proto_rawDesc = []byte
0x72, 0x65, 0x6d, 0x6f, 0x74, 0x65, 0x2e, 0x65, 0x78, 0x65, 0x63, 0x75, 0x74, 0x69, 0x6f, 0x6e,
0x2e, 0x76, 0x32, 0x2e, 0x42, 0x61, 0x74, 0x63, 0x68, 0x52, 0x65, 0x61, 0x64, 0x42, 0x6c, 0x6f,
0x62, 0x73, 0x52, 0x65, 0x73, 0x70, 0x6f, 0x6e, 0x73, 0x65, 0x22, 0x31, 0x82, 0xd3, 0xe4, 0x93,
- 0x02, 0x2b, 0x22, 0x26, 0x2f, 0x76, 0x32, 0x2f, 0x7b, 0x69, 0x6e, 0x73, 0x74, 0x61, 0x6e, 0x63,
- 0x65, 0x5f, 0x6e, 0x61, 0x6d, 0x65, 0x3d, 0x2a, 0x2a, 0x7d, 0x2f, 0x62, 0x6c, 0x6f, 0x62, 0x73,
- 0x3a, 0x62, 0x61, 0x74, 0x63, 0x68, 0x52, 0x65, 0x61, 0x64, 0x3a, 0x01, 0x2a, 0x12, 0xc8, 0x01,
+ 0x02, 0x2b, 0x3a, 0x01, 0x2a, 0x22, 0x26, 0x2f, 0x76, 0x32, 0x2f, 0x7b, 0x69, 0x6e, 0x73, 0x74,
+ 0x61, 0x6e, 0x63, 0x65, 0x5f, 0x6e, 0x61, 0x6d, 0x65, 0x3d, 0x2a, 0x2a, 0x7d, 0x2f, 0x62, 0x6c,
+ 0x6f, 0x62, 0x73, 0x3a, 0x62, 0x61, 0x74, 0x63, 0x68, 0x52, 0x65, 0x61, 0x64, 0x12, 0xc8, 0x01,
0x0a, 0x07, 0x47, 0x65, 0x74, 0x54, 0x72, 0x65, 0x65, 0x12, 0x2f, 0x2e, 0x62, 0x75, 0x69, 0x6c,
0x64, 0x2e, 0x62, 0x61, 0x7a, 0x65, 0x6c, 0x2e, 0x72, 0x65, 0x6d, 0x6f, 0x74, 0x65, 0x2e, 0x65,
0x78, 0x65, 0x63, 0x75, 0x74, 0x69, 0x6f, 0x6e, 0x2e, 0x76, 0x32, 0x2e, 0x47, 0x65, 0x74, 0x54,
@@ -5711,10 +5712,10 @@ var file_build_bazel_remote_execution_v2_remote_execution_proto_goTypes = []inte
(*BatchUpdateBlobsResponse_Response)(nil), // 53: build.bazel.remote.execution.v2.BatchUpdateBlobsResponse.Response
(*BatchReadBlobsResponse_Response)(nil), // 54: build.bazel.remote.execution.v2.BatchReadBlobsResponse.Response
(*PriorityCapabilities_PriorityRange)(nil), // 55: build.bazel.remote.execution.v2.PriorityCapabilities.PriorityRange
- (*duration.Duration)(nil), // 56: google.protobuf.Duration
- (*timestamp.Timestamp)(nil), // 57: google.protobuf.Timestamp
- (*wrappers.UInt32Value)(nil), // 58: google.protobuf.UInt32Value
- (*any.Any)(nil), // 59: google.protobuf.Any
+ (*durationpb.Duration)(nil), // 56: google.protobuf.Duration
+ (*timestamppb.Timestamp)(nil), // 57: google.protobuf.Timestamp
+ (*wrapperspb.UInt32Value)(nil), // 58: google.protobuf.UInt32Value
+ (*anypb.Any)(nil), // 59: google.protobuf.Any
(*status.Status)(nil), // 60: google.rpc.Status
(*semver.SemVer)(nil), // 61: build.bazel.semver.SemVer
(*longrunning.Operation)(nil), // 62: google.longrunning.Operation
@@ -6536,21 +6537,21 @@ type ExecutionClient interface {
// server MUST NOT set the `error` field of the `Operation` proto.
// The possible errors include:
//
- // * `INVALID_ARGUMENT`: One or more arguments are invalid.
- // * `FAILED_PRECONDITION`: One or more errors occurred in setting up the
- // action requested, such as a missing input or command or no worker being
- // available. The client may be able to fix the errors and retry.
- // * `RESOURCE_EXHAUSTED`: There is insufficient quota of some resource to run
- // the action.
- // * `UNAVAILABLE`: Due to a transient condition, such as all workers being
- // occupied (and the server does not support a queue), the action could not
- // be started. The client should retry.
- // * `INTERNAL`: An internal error occurred in the execution engine or the
- // worker.
- // * `DEADLINE_EXCEEDED`: The execution timed out.
- // * `CANCELLED`: The operation was cancelled by the client. This status is
- // only possible if the server implements the Operations API CancelOperation
- // method, and it was called for the current execution.
+ // - `INVALID_ARGUMENT`: One or more arguments are invalid.
+ // - `FAILED_PRECONDITION`: One or more errors occurred in setting up the
+ // action requested, such as a missing input or command or no worker being
+ // available. The client may be able to fix the errors and retry.
+ // - `RESOURCE_EXHAUSTED`: There is insufficient quota of some resource to run
+ // the action.
+ // - `UNAVAILABLE`: Due to a transient condition, such as all workers being
+ // occupied (and the server does not support a queue), the action could not
+ // be started. The client should retry.
+ // - `INTERNAL`: An internal error occurred in the execution engine or the
+ // worker.
+ // - `DEADLINE_EXCEEDED`: The execution timed out.
+ // - `CANCELLED`: The operation was cancelled by the client. This status is
+ // only possible if the server implements the Operations API CancelOperation
+ // method, and it was called for the current execution.
//
// In the case of a missing input or command, the server SHOULD additionally
// send a [PreconditionFailure][google.rpc.PreconditionFailure] error detail
@@ -6574,6 +6575,14 @@ type ExecutionClient interface {
// operation completes, and then respond with the completed operation. The
// server MAY choose to stream additional updates as execution progresses,
// such as to provide an update as to the state of the execution.
+ //
+ // In addition to the cases describe for Execute, the WaitExecution method
+ // may fail as follows:
+ //
+ // - `NOT_FOUND`: The operation no longer exists due to any of a transient
+ // condition, an unknown operation name, or if the server implements the
+ // Operations API DeleteOperation method and it was called for the current
+ // execution. The client should call `Execute` to retry.
WaitExecution(ctx context.Context, in *WaitExecutionRequest, opts ...grpc.CallOption) (Execution_WaitExecutionClient, error)
}
@@ -6693,21 +6702,21 @@ type ExecutionServer interface {
// server MUST NOT set the `error` field of the `Operation` proto.
// The possible errors include:
//
- // * `INVALID_ARGUMENT`: One or more arguments are invalid.
- // * `FAILED_PRECONDITION`: One or more errors occurred in setting up the
- // action requested, such as a missing input or command or no worker being
- // available. The client may be able to fix the errors and retry.
- // * `RESOURCE_EXHAUSTED`: There is insufficient quota of some resource to run
- // the action.
- // * `UNAVAILABLE`: Due to a transient condition, such as all workers being
- // occupied (and the server does not support a queue), the action could not
- // be started. The client should retry.
- // * `INTERNAL`: An internal error occurred in the execution engine or the
- // worker.
- // * `DEADLINE_EXCEEDED`: The execution timed out.
- // * `CANCELLED`: The operation was cancelled by the client. This status is
- // only possible if the server implements the Operations API CancelOperation
- // method, and it was called for the current execution.
+ // - `INVALID_ARGUMENT`: One or more arguments are invalid.
+ // - `FAILED_PRECONDITION`: One or more errors occurred in setting up the
+ // action requested, such as a missing input or command or no worker being
+ // available. The client may be able to fix the errors and retry.
+ // - `RESOURCE_EXHAUSTED`: There is insufficient quota of some resource to run
+ // the action.
+ // - `UNAVAILABLE`: Due to a transient condition, such as all workers being
+ // occupied (and the server does not support a queue), the action could not
+ // be started. The client should retry.
+ // - `INTERNAL`: An internal error occurred in the execution engine or the
+ // worker.
+ // - `DEADLINE_EXCEEDED`: The execution timed out.
+ // - `CANCELLED`: The operation was cancelled by the client. This status is
+ // only possible if the server implements the Operations API CancelOperation
+ // method, and it was called for the current execution.
//
// In the case of a missing input or command, the server SHOULD additionally
// send a [PreconditionFailure][google.rpc.PreconditionFailure] error detail
@@ -6731,6 +6740,14 @@ type ExecutionServer interface {
// operation completes, and then respond with the completed operation. The
// server MAY choose to stream additional updates as execution progresses,
// such as to provide an update as to the state of the execution.
+ //
+ // In addition to the cases describe for Execute, the WaitExecution method
+ // may fail as follows:
+ //
+ // - `NOT_FOUND`: The operation no longer exists due to any of a transient
+ // condition, an unknown operation name, or if the server implements the
+ // Operations API DeleteOperation method and it was called for the current
+ // execution. The client should call `Execute` to retry.
WaitExecution(*WaitExecutionRequest, Execution_WaitExecutionServer) error
}
@@ -6841,11 +6858,11 @@ type ActionCacheClient interface {
//
// Errors:
//
- // * `INVALID_ARGUMENT`: One or more arguments are invalid.
- // * `FAILED_PRECONDITION`: One or more errors occurred in updating the
- // action result, such as a missing command or action.
- // * `RESOURCE_EXHAUSTED`: There is insufficient storage space to add the
- // entry to the cache.
+ // - `INVALID_ARGUMENT`: One or more arguments are invalid.
+ // - `FAILED_PRECONDITION`: One or more errors occurred in updating the
+ // action result, such as a missing command or action.
+ // - `RESOURCE_EXHAUSTED`: There is insufficient storage space to add the
+ // entry to the cache.
UpdateActionResult(ctx context.Context, in *UpdateActionResultRequest, opts ...grpc.CallOption) (*ActionResult, error)
}
@@ -6904,11 +6921,11 @@ type ActionCacheServer interface {
//
// Errors:
//
- // * `INVALID_ARGUMENT`: One or more arguments are invalid.
- // * `FAILED_PRECONDITION`: One or more errors occurred in updating the
- // action result, such as a missing command or action.
- // * `RESOURCE_EXHAUSTED`: There is insufficient storage space to add the
- // entry to the cache.
+ // - `INVALID_ARGUMENT`: One or more arguments are invalid.
+ // - `FAILED_PRECONDITION`: One or more errors occurred in updating the
+ // action result, such as a missing command or action.
+ // - `RESOURCE_EXHAUSTED`: There is insufficient storage space to add the
+ // entry to the cache.
UpdateActionResult(context.Context, *UpdateActionResultRequest) (*ActionResult, error)
}
@@ -7009,8 +7026,8 @@ type ContentAddressableStorageClient interface {
//
// Errors:
//
- // * `INVALID_ARGUMENT`: The client attempted to upload more than the
- // server supported limit.
+ // - `INVALID_ARGUMENT`: The client attempted to upload more than the
+ // server supported limit.
//
// Individual requests may return the following errors, additionally:
//
@@ -7034,8 +7051,8 @@ type ContentAddressableStorageClient interface {
//
// Errors:
//
- // * `INVALID_ARGUMENT`: The client attempted to read more than the
- // server supported limit.
+ // - `INVALID_ARGUMENT`: The client attempted to read more than the
+ // server supported limit.
//
// Every error on individual read will be returned in the corresponding digest
// status.
@@ -7160,8 +7177,8 @@ type ContentAddressableStorageServer interface {
//
// Errors:
//
- // * `INVALID_ARGUMENT`: The client attempted to upload more than the
- // server supported limit.
+ // - `INVALID_ARGUMENT`: The client attempted to upload more than the
+ // server supported limit.
//
// Individual requests may return the following errors, additionally:
//
@@ -7185,8 +7202,8 @@ type ContentAddressableStorageServer interface {
//
// Errors:
//
- // * `INVALID_ARGUMENT`: The client attempted to read more than the
- // server supported limit.
+ // - `INVALID_ARGUMENT`: The client attempted to read more than the
+ // server supported limit.
//
// Every error on individual read will be returned in the corresponding digest
// status.
@@ -7348,10 +7365,10 @@ type CapabilitiesClient interface {
// remote endpoint.
// Only the capabilities of the services supported by the endpoint will
// be returned:
- // * Execution + CAS + Action Cache endpoints should return both
- // CacheCapabilities and ExecutionCapabilities.
- // * Execution only endpoints should return ExecutionCapabilities.
- // * CAS + Action Cache only endpoints should return CacheCapabilities.
+ // - Execution + CAS + Action Cache endpoints should return both
+ // CacheCapabilities and ExecutionCapabilities.
+ // - Execution only endpoints should return ExecutionCapabilities.
+ // - CAS + Action Cache only endpoints should return CacheCapabilities.
//
// There are no method-specific errors.
GetCapabilities(ctx context.Context, in *GetCapabilitiesRequest, opts ...grpc.CallOption) (*ServerCapabilities, error)
@@ -7380,10 +7397,10 @@ type CapabilitiesServer interface {
// remote endpoint.
// Only the capabilities of the services supported by the endpoint will
// be returned:
- // * Execution + CAS + Action Cache endpoints should return both
- // CacheCapabilities and ExecutionCapabilities.
- // * Execution only endpoints should return ExecutionCapabilities.
- // * CAS + Action Cache only endpoints should return CacheCapabilities.
+ // - Execution + CAS + Action Cache endpoints should return both
+ // CacheCapabilities and ExecutionCapabilities.
+ // - Execution only endpoints should return ExecutionCapabilities.
+ // - CAS + Action Cache only endpoints should return CacheCapabilities.
//
// There are no method-specific errors.
GetCapabilities(context.Context, *GetCapabilitiesRequest) (*ServerCapabilities, error)
diff --git a/external/BUILD.googleapis b/external/BUILD.googleapis
deleted file mode 100644
index 5225788..0000000
--- a/external/BUILD.googleapis
+++ /dev/null
@@ -1,78 +0,0 @@
-package(default_visibility = ["//visibility:public"])
-
-load("@rules_proto//proto:defs.bzl", "proto_library")
-load("@com_github_grpc_grpc//bazel:cc_grpc_library.bzl", "cc_grpc_library")
-
-licenses(["notice"])
-
-proto_library(
- name = "google_api_annotations_proto",
- srcs = ["google/api/annotations.proto"],
- deps = [
- ":google_api_http_proto",
- "@com_google_protobuf//:descriptor_proto",
- ],
-)
-
-cc_grpc_library(
- name = "google_api_annotations_cc_proto",
- srcs = ["google/api/annotations.proto"],
- deps = [":google_api_http_cc_proto"],
- proto_only = False,
- well_known_protos = True,
- use_external = False,
-)
-
-proto_library(
- name = "google_api_http_proto",
- srcs = ["google/api/http.proto"],
-)
-
-cc_grpc_library(
- name = "google_api_http_cc_proto",
- srcs = ["google/api/http.proto"],
- deps = [],
- proto_only = False,
- well_known_protos = True,
- use_external = False,
-)
-
-proto_library(
- name = "google_longrunning_operations_proto",
- srcs = ["google/longrunning/operations.proto"],
- deps = [
- ":google_api_annotations_proto",
- ":google_api_http_proto",
- ":google_rpc_status_proto",
- "@com_google_protobuf//:any_proto",
- "@com_google_protobuf//:empty_proto",
- ],
-)
-
-cc_grpc_library(
- name = "google_longrunning_operations_cc_proto",
- srcs = ["google/longrunning/operations.proto"],
- deps = [
- ":google_api_annotations_cc_proto",
- ":google_api_http_cc_proto",
- ":google_rpc_status_cc_proto",
- ],
- proto_only = False,
- well_known_protos = True,
- use_external = False,
-)
-
-proto_library(
- name = "google_rpc_status_proto",
- srcs = ["google/rpc/status.proto"],
- deps = ["@com_google_protobuf//:any_proto"],
-)
-
-cc_grpc_library(
- name = "google_rpc_status_cc_proto",
- srcs = ["google/rpc/status.proto"],
- deps = [],
- proto_only = False,
- well_known_protos = True,
- use_external = False,
-)
diff --git a/remote_apis_deps.bzl b/remote_apis_deps.bzl
index a3a2383..c828a2e 100644
--- a/remote_apis_deps.bzl
+++ b/remote_apis_deps.bzl
@@ -9,10 +9,22 @@ def _maybe(repo_rule, name, **kwargs):
def remote_apis_go_deps():
"""Load dependencies needed to depend on RE API for Go"""
- go_rules_dependencies()
- go_download_sdk(name = "go_sdk", version = "1.16.4")
+ go_download_sdk(name = "go_sdk", version = "1.20.6")
go_register_toolchains()
- gazelle_dependencies()
+
+ # The version of this repo needs to be in sync with @googleapis
+ go_repository(
+ name = "org_golang_google_genproto",
+ build_extra_args = ["-exclude=vendor"],
+ build_file_generation = "on",
+ build_file_proto_mode = "disable_global",
+ importpath = "google.golang.org/genproto",
+ sum = "h1:S9GbmC1iCgvbLyAokVCwiO6tVIrU9Y7c5oMx1V/ki/Y=",
+ version = "v0.0.0-20221024183307-1bc688fe9f3e",
+ )
+
+ go_rules_dependencies()
+ gazelle_dependencies(go_sdk = "go_sdk")
_maybe(
go_repository,
name = "com_github_golang_protobuf",